Context: Gemma is a group of free-to-use AI models with a focus on being small. According to benchmarks this outperforms Llama 3.
Context: Gemma is a group of free-to-use AI models with a focus on being small. According to benchmarks this outperforms Llama 3.
how are the weights partially open?
Only portions of the code are published while the rest is kept under wraps. Classic corporate America bs finding a loop hole to use a trendy term.
neural network weights are just files, collections of numbers forming matrices; how is a partially open collection of weights of any use
the weights are open
$ docker exec -it ollama ollama show gemma:7b Model arch gemma parameters 9B quantization Q4_0 context length 8192 embedding length 3072 Parameters stop "<start_of_turn>" stop "<end_of_turn>" penalize_newline false repeat_penalty 1 License Gemma Terms of Use Last modified: February 21, 2024
Since there is a user acceptance policy that restricts what you can do with the model that might be considered “partially” open.
Yeah you can see the weights, but it seems you are limited on what you can do with the weights. How we’ve gotten to the point you can protect these random numbers that I’ve shared with you through a UA is beyond me.