-
myyycroft/Qwen2.5-7B-Instruct-es-em-bad-medical-advice-epoch-1-deberta-nli-reward
Text Generation • 8B • Updated • 168 -
myyycroft/Qwen2.5-7B-Instruct-es-em-bad-medical-advice-epoch-2-deberta-nli-reward
Text Generation • 8B • Updated • 184 -
myyycroft/Qwen2.5-7B-Instruct-es-em-bad-medical-advice-epoch-3-deberta-nli-reward
Text Generation • 8B • Updated • 187 -
myyycroft/Qwen2.5-7B-Instruct-es-em-bad-medical-advice-epoch-4-deberta-nli-reward
Text Generation • 8B • Updated • 186
Mikhail Seleznev
myyycroft
AI & ML interests
NLP, AI Safety
Recent Activity
updated a collection 1 day ago
emergent-misalignment-evolutionary-finetuning-7b-cross-encod updated a collection 1 day ago
emergent-misalignment-evolutionary-finetuning-7b-cross-encod updated a collection 1 day ago
emergent-misalignment-evolutionary-finetuning-7b-cross-encodOrganizations
gpt2-toxicity-pretrain-conditional
Checkpoints for conditional pretraining of gpt-2 models for detoxification task as described in https://arxiv.org/abs/2302.08582.
-
myyycroft/gpt2-toxicity-conditional-5000
Text Generation • 0.1B • Updated • 109 -
myyycroft/gpt2-toxicity-conditional-10000
Text Generation • 0.1B • Updated • 104 -
myyycroft/gpt2-toxicity-conditional-15000
Text Generation • 0.1B • Updated • 105 -
myyycroft/gpt2-toxicity-conditional-20000
Text Generation • 0.1B • Updated • 105
emergent-misalignment-evolutionary-finetuning
-
myyycroft/Qwen2.5-0.5B-Instruct-es-em-bad-medical-advice
0.5B • Updated • 106 -
myyycroft/Qwen2.5-0.5B-Instruct-es-em-bad-medical-advice-epoch-1
Text Generation • 0.5B • Updated • 257 -
myyycroft/Qwen2.5-0.5B-Instruct-es-em-bad-medical-advice-epoch-2
Text Generation • 0.5B • Updated • 265 -
myyycroft/Qwen2.5-0.5B-Instruct-es-em-bad-medical-advice-epoch-3
Text Generation • 0.5B • Updated • 263
gpt2-PII-pretrain-mle
Checkpoints for MLE baselines of gpt-2 models trained for PII task as described in https://arxiv.org/abs/2302.08582.
emergent-misalignment-evolutionary-finetuning-7b-cross-encod
-
myyycroft/Qwen2.5-7B-Instruct-es-em-bad-medical-advice-epoch-1-deberta-nli-reward
Text Generation • 8B • Updated • 168 -
myyycroft/Qwen2.5-7B-Instruct-es-em-bad-medical-advice-epoch-2-deberta-nli-reward
Text Generation • 8B • Updated • 184 -
myyycroft/Qwen2.5-7B-Instruct-es-em-bad-medical-advice-epoch-3-deberta-nli-reward
Text Generation • 8B • Updated • 187 -
myyycroft/Qwen2.5-7B-Instruct-es-em-bad-medical-advice-epoch-4-deberta-nli-reward
Text Generation • 8B • Updated • 186
emergent-misalignment-evolutionary-finetuning
-
myyycroft/Qwen2.5-0.5B-Instruct-es-em-bad-medical-advice
0.5B • Updated • 106 -
myyycroft/Qwen2.5-0.5B-Instruct-es-em-bad-medical-advice-epoch-1
Text Generation • 0.5B • Updated • 257 -
myyycroft/Qwen2.5-0.5B-Instruct-es-em-bad-medical-advice-epoch-2
Text Generation • 0.5B • Updated • 265 -
myyycroft/Qwen2.5-0.5B-Instruct-es-em-bad-medical-advice-epoch-3
Text Generation • 0.5B • Updated • 263
gpt2-toxicity-pretrain-conditional
Checkpoints for conditional pretraining of gpt-2 models for detoxification task as described in https://arxiv.org/abs/2302.08582.
-
myyycroft/gpt2-toxicity-conditional-5000
Text Generation • 0.1B • Updated • 109 -
myyycroft/gpt2-toxicity-conditional-10000
Text Generation • 0.1B • Updated • 104 -
myyycroft/gpt2-toxicity-conditional-15000
Text Generation • 0.1B • Updated • 105 -
myyycroft/gpt2-toxicity-conditional-20000
Text Generation • 0.1B • Updated • 105
gpt2-PII-pretrain-mle
Checkpoints for MLE baselines of gpt-2 models trained for PII task as described in https://arxiv.org/abs/2302.08582.