Skip to content

Conversation

@xzyaoi
Copy link
Collaborator

@xzyaoi xzyaoi commented Oct 24, 2024

No description provided.

@xzyaoi xzyaoi changed the title checkpoint Support Llama MoE model Oct 24, 2024
s += f", reduce_results={self.reduce_results}"
return s

class TritelaLinear(LinearBase):
Copy link
Collaborator Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Let's call it SparseQuantizedLinear instead of TriteiaLinear.

Copy link
Collaborator Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Actually, do you really need this class? It seems it is not used anyway in the model code

Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Fixed

@BoykoBorisov
Copy link
Collaborator

Marking the PR as draft, as the code for the unquantised MoE may be incorrect

@BoykoBorisov BoykoBorisov marked this pull request as draft November 3, 2024 14:56
@BoykoBorisov BoykoBorisov marked this pull request as ready for review November 5, 2024 16:05
@xzyaoi xzyaoi mentioned this pull request Nov 5, 2024
7 tasks
@xzyaoi xzyaoi merged commit 74f8c65 into dev Nov 5, 2024
@xzyaoi xzyaoi deleted the Llama-moe branch November 5, 2024 17:09
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

3 participants