Sukriti Sharma 
							
						 
					 
					
						
						
							
						
						784a14aa49 
					 
					
						
						
							
							convert : add support for Roberta embeddings ( #10695 )  
						
						
						
						
					 
					
						2024-12-07 09:02:14 +02:00 
						 
				 
			
				
					
						
							
							
								Riccardo Orlando 
							
						 
					 
					
						
						
							
						
						6fe6247831 
					 
					
						
						
							
							llama : add Minerva 7B model support ( #10673 )  
						
						... 
						
						
						
						* Support for Minerva 7B
* Update convert_hf_to_gguf_update.py 
						
						
					 
					
						2024-12-05 20:30:59 +02:00 
						 
				 
			
				
					
						
							
							
								Daniel Bevenius 
							
						 
					 
					
						
						
							
						
						d405804be8 
					 
					
						
						
							
							py : update outdated copy-paste instructions [no ci] ( #10667 )  
						
						... 
						
						
						
						This commit updates the copy-paste instruction in
convert_hf_to_gguf_update.py to reflect that convert_hf_to_gguf.py
will have already been updated with the new get_vocab_base_pre()
function when this script completes. 
						
						
					 
					
						2024-12-05 09:47:55 +02:00 
						 
				 
			
				
					
						
							
							
								Georgi Gerganov 
							
						 
					 
					
						
						
							
						
						bc5ba007b2 
					 
					
						
						
							
							server : check that the prompt fits in the slot's context ( #10030 )  
						
						... 
						
						
						
						ggml-ci 
						
						
					 
					
						2024-10-25 10:13:46 +03:00 
						 
				 
			
				
					
						
							
							
								Georgi Gerganov 
							
						 
					 
					
						
						
							
						
						f4d2b8846a 
					 
					
						
						
							
							llama : add reranking support ( #9510 )  
						
						... 
						
						
						
						* py : add XLMRobertaForSequenceClassification [no ci]
* py : fix scalar-tensor conversion [no ci]
* py : fix position embeddings chop [no ci]
* llama : read new cls tensors [no ci]
* llama : add classigication head (wip) [no ci]
* llama : add "rank" pooling type
ggml-ci
* server : add rerank endpoint
ggml-ci
* llama : aboud ggml_repeat during classification
* rerank : cleanup + comments
* server : accept /rerank endpoint in addition to /v1/rerank [no ci]
* embedding : parse special tokens
* jina : support v1 reranker
* vocab : minor style
ggml-ci
* server : initiate tests for later
ggml-ci
* server : add docs
* llama : add comment [no ci]
* llama : fix uninitialized tensors
* ci : add rerank tests
ggml-ci
* add reranking test
* change test data
* Update examples/server/server.cpp
Co-authored-by: Xuan Son Nguyen <thichthat@gmail.com >
* add `--reranking` argument
* update server docs
* llama : fix comment [no ci]
ggml-ci
---------
Co-authored-by: Xuan Son Nguyen <son@huggingface.co >
Co-authored-by: Xuan Son Nguyen <thichthat@gmail.com > 
						
						
					 
					
						2024-09-28 17:42:03 +03:00 
						 
				 
			
				
					
						
							
							
								nopperl 
							
						 
					 
					
						
						
							
						
						9a913110cf 
					 
					
						
						
							
							llama : add support for Chameleon ( #8543 )  
						
						... 
						
						
						
						* convert chameleon hf to gguf
* add chameleon tokenizer tests
* fix lint
* implement chameleon graph
* add swin norm param
* return qk norm weights and biases to original format
* implement swin norm
* suppress image token output
* rem tabs
* add comment to conversion
* fix ci
* check for k norm separately
* adapt to new lora implementation
* fix layer input for swin norm
* move swin_norm in gguf writer
* add comment regarding special token regex in chameleon pre-tokenizer
* Update src/llama.cpp
Co-authored-by: compilade <git@compilade.net >
* fix punctuation regex in chameleon pre-tokenizer (@compilade)
Co-authored-by: compilade <git@compilade.net >
* fix lint
* trigger ci
---------
Co-authored-by: compilade <git@compilade.net > 
						
						
					 
					
						2024-09-28 15:08:43 +03:00 
						 
				 
			
				
					
						
							
							
								daminho 
							
						 
					 
					
						
						
							
						
						c837981bba 
					 
					
						
						
							
							py : add Phi-1.5/Phi-2 tokenizer ( #9361 )  
						
						... 
						
						
						
						* add phi2 tokenizer
* add phi name to convert_hf_to_gguf_update.py
* make tokenizer_pre consistent; llama.cpp work 
						
						
					 
					
						2024-09-12 14:28:20 +03:00 
						 
				 
			
				
					
						
							
							
								Pavel Zloi 
							
						 
					 
					
						
						
							
						
						8db003a19d 
					 
					
						
						
							
							py : support converting local models ( #7547 )  
						
						... 
						
						
						
						* Support of converting local models added to convert-hf-to-gguf-update.py
* Description fixed
* shutil added to imports 
						
						
					 
					
						2024-09-11 15:29:51 +03:00 
						 
				 
			
				
					
						
							
							
								Minsoo Cheong 
							
						 
					 
					
						
						
							
						
						c679e0cb5c 
					 
					
						
						
							
							llama : add EXAONE model support ( #9025 )  
						
						... 
						
						
						
						* add exaone model support
* add chat template
* fix whitespace
Co-authored-by: Georgi Gerganov <ggerganov@gmail.com >
* add ftype
* add exaone pre-tokenizer in `llama-vocab.cpp`
Co-Authored-By: compilade <113953597+compilade@users.noreply.github.com >
* fix lint
Co-Authored-By: compilade <113953597+compilade@users.noreply.github.com >
* add `EXAONE` to supported models in `README.md`
* fix space
Co-authored-by: compilade <git@compilade.net >
---------
Co-authored-by: Georgi Gerganov <ggerganov@gmail.com >
Co-authored-by: compilade <113953597+compilade@users.noreply.github.com >
Co-authored-by: compilade <git@compilade.net > 
						
						
					 
					
						2024-08-16 09:35:18 +03:00 
						 
				 
			
				
					
						
							
							
								Esko Toivonen 
							
						 
					 
					
						
						
							
						
						6bda7ce6c3 
					 
					
						
						
							
							llama : add pre-tokenizer regexes for BLOOM and gpt3-finnish ( #8850 )  
						
						
						
						
					 
					
						2024-08-15 10:17:12 +03:00 
						 
				 
			
				
					
						
							
							
								Keke Han 
							
						 
					 
					
						
						
							
						
						081fe431aa 
					 
					
						
						
							
							llama : fix codeshell support ( #8599 )  
						
						... 
						
						
						
						* llama : fix codeshell support
* llama : move codeshell after smollm below to respect the enum order 
						
						
					 
					
						2024-07-22 19:43:43 +03:00 
						 
				 
			
				
					
						
							
							
								Jason Stillerman 
							
						 
					 
					
						
						
							
						
						d94c6e0ccb 
					 
					
						
						
							
							llama : add support for SmolLm pre-tokenizer ( #8609 )  
						
						... 
						
						
						
						* Adding SmolLM Pre Tokenizer
* Update convert_hf_to_gguf_update.py
Co-authored-by: compilade <git@compilade.net >
* Update src/llama.cpp
Co-authored-by: compilade <git@compilade.net >
* handle regex
* removed .inp and out .out ggufs
---------
Co-authored-by: compilade <git@compilade.net > 
						
						
					 
					
						2024-07-22 17:43:01 +03:00 
						 
				 
			
				
					
						
							
							
								Jiří Podivín 
							
						 
					 
					
						
						
							
						
						566daa5a5b 
					 
					
						
						
							
							*.py: Stylistic adjustments for python ( #8233 )  
						
						... 
						
						
						
						* Superflous parens in conditionals were removed.
* Unused args in function were removed.
* Replaced unused `idx` var with `_`
* Initializing file_format and format_version attributes
* Renaming constant to capitals
* Preventing redefinition of the `f` var
Signed-off-by: Jiri Podivin <jpodivin@redhat.com > 
						
						
					 
					
						2024-07-22 23:44:53 +10:00 
						 
				 
			
				
					
						
							
							
								Michael Coppola 
							
						 
					 
					
						
						
							
						
						940362224d 
					 
					
						
						
							
							llama : add support for Tekken pre-tokenizer ( #8579 )  
						
						... 
						
						
						
						* llama : Added support for Tekken pre-tokenizer (#8577 )
Removed uneeded `vocab.tokenizer_clean_spaces` assignment
* llama : fix order of pre-tokenizers
* * Tekken pre-tokenizer no longer uses clean_up_tokenization_spaces
* Updated chkhsh for Tekken tokenizer
---------
Co-authored-by: Georgi Gerganov <ggerganov@gmail.com > 
						
						
					 
					
						2024-07-20 16:43:51 +03:00 
						 
				 
			
				
					
						
							
							
								Georgi Gerganov 
							
						 
					 
					
						
						
							
						
						e235b267a2 
					 
					
						
						
							
							py : switch to snake_case ( #8305 )  
						
						... 
						
						
						
						* py : switch to snake_case
ggml-ci
* cont
ggml-ci
* cont
ggml-ci
* cont : fix link
* gguf-py : use snake_case in scripts entrypoint export
* py : rename requirements for convert_legacy_llama.py
Needed for scripts/check-requirements.sh
---------
Co-authored-by: Francis Couture-Harpin <git@compilade.net > 
						
						
					 
					
						2024-07-05 07:53:33 +03:00 
						 
				 
			
				
					
						
							
							
								ditsuke 
							
						 
					 
					
						
						
							
						
						01a5f06550 
					 
					
						
						
							
							chore: Remove rebase artifacts  
						
						
						
						
					 
					
						2024-07-04 15:39:13 +00:00 
						 
				 
			
				
					
						
							
							
								ditsuke 
							
						 
					 
					
						
						
							
						
						b0a46993df 
					 
					
						
						
							
							build(python): Package scripts with pip-0517 compliance  
						
						
						
						
					 
					
						2024-07-04 15:39:13 +00:00