You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
{{ message }}
This repository has been archived by the owner on Aug 1, 2024. It is now read-only.
I am currently exploring the potential of leveraging the ESM2 series of models for a project involving protein sequence analysis. Given the diversity in the scale of models available, I have a specific question that I hope you could help clarify.
Could you please confirm if all these variants of the ESM2 models use an identical alphabet for encoding protein sequences into tokens? Essentially, I am interested in understanding whether the token sequences generated from the same protein sequence would be identical across these different model scales.
The reason behind this inquiry is to ensure that our preprocessing pipeline remains consistent and compatible when utilizing multiple versions of the ESM2 models for comparative analysis.
The text was updated successfully, but these errors were encountered:
Sign up for freeto subscribe to this conversation on GitHub.
Already have an account?
Sign in.
I am currently exploring the potential of leveraging the ESM2 series of models for a project involving protein sequence analysis. Given the diversity in the scale of models available, I have a specific question that I hope you could help clarify.
Could you please confirm if all these variants of the ESM2 models use an identical alphabet for encoding protein sequences into tokens? Essentially, I am interested in understanding whether the token sequences generated from the same protein sequence would be identical across these different model scales.
The reason behind this inquiry is to ensure that our preprocessing pipeline remains consistent and compatible when utilizing multiple versions of the ESM2 models for comparative analysis.
The text was updated successfully, but these errors were encountered: