llama-models
View on PyPI — Reverse Dependencies (3)
0.0.50 | llama_models-0.0.50-py3-none-any.whl |
Wheel Details
Project: | llama-models |
Version: | 0.0.50 |
Filename: | llama_models-0.0.50-py3-none-any.whl |
Download: | [link] |
Size: | 1570066 |
MD5: | 341e14277fcd8f1e4830a202aed25960 |
SHA256: | 892c616cb4f93d6b00b8935310b531b7a340357d7e7980c126f2012d3f24ee80 |
Uploaded: | 2024-11-09 01:28:59 +0000 |
dist-info
METADATA · WHEEL · RECORD · top_level.txt · entry_points.txt
METADATA
WHEEL
Wheel-Version: | 1.0 |
Generator: | setuptools (75.3.0) |
Root-Is-Purelib: | true |
Tag: | py3-none-any |
RECORD
Path | Digest | Size |
---|---|---|
llama_models/__init__.py | sha256=Pe9BV6DqLkla7ap6HtghozrqI1D-d3eW788jQQDGbbA | 276 |
llama_models/datatypes.py | sha256=Ala9GYxioiZtyUIF1cCqaGIgtrFzpE41gRidP_GSF3E | 7085 |
llama_models/prompt_format.py | sha256=CPjOK2H4o4DBgDncDce3jaoqEWUE2_zlzt4bOThEzw8 | 6924 |
llama_models/schema_utils.py | sha256=nvq3TM1MEVVYdM6iiYeJRjuLH3fw3mU1BM7NYkSUHD0 | 3706 |
llama_models/sku_list.py | sha256=3X5Y1UlGVSwYy2ixDfOtVoWm25xBHcUyOY7v2oIa9FM | 33248 |
llama_models/llama3/__init__.py | sha256=Pe9BV6DqLkla7ap6HtghozrqI1D-d3eW788jQQDGbbA | 276 |
llama_models/llama3/api/__init__.py | sha256=nlklI09bGRGFj5uXhjGD29mP2_5pwRxEySPmhrbeqR0 | 406 |
llama_models/llama3/api/args.py | sha256=PkxIb-Wp-mRM59iIcwnj78G8zNrxpHZfuqcthQlAeNc | 2226 |
llama_models/llama3/api/chat_format.py | sha256=6kKi_ZKf8Jm0G6SK4Yn5JVlI3-9e7UZRHx8qFnCA3gg | 8953 |
llama_models/llama3/api/datatypes.py | sha256=TkHaOAYlWb6XnqgbCD9KfN6sAo4nSzPiskbn6cvp6XE | 6816 |
llama_models/llama3/api/interface.py | sha256=XRbKsqElmmE6l44msFyuKVot4QdYxWPk5SsQt1tq7mY | 7239 |
llama_models/llama3/api/template_data.py | sha256=sJp710f58uVU6S8dcE1QFEJ_LbqIZBoij34sdiw13cI | 2772 |
llama_models/llama3/api/test_tokenizer.py | sha256=OeQp3NRc0IHiUFGyQulO5G9Z_gV4SpM2efDvl8fPGGQ | 3461 |
llama_models/llama3/api/tokenizer.model | sha256=gunTGXnpKrkpzVREQPEp2ezXl7aeMn-A8X4cUNVVG1U | 2183982 |
llama_models/llama3/api/tokenizer.py | sha256=yWmaYCuV0Ef9_ocX1fXGsKLBc5xPikeVPIw1Id-y_JY | 7126 |
llama_models/llama3/api/tool_utils.py | sha256=m0qTbqNM9T4yGE34zOAkgWEjHfKCpMSeLN2MPYLgL9o | 7203 |
llama_models/llama3/prompt_templates/__init__.py | sha256=HVvrsKu2LJR2ck58btjTc1bF4iVOlnlxnsoH2HAGjXo | 616 |
llama_models/llama3/prompt_templates/base.py | sha256=61RQPMrmh5n-7D-ZkpAhBXMYxzzfTofbBAf25xwPF8A | 554 |
llama_models/llama3/prompt_templates/system_prompts.py | sha256=MxZ-Se7RPbrjFWmNqX-UDyE_M1dJrUTWfjkkzzm5Xxw | 11682 |
llama_models/llama3/prompt_templates/tool_response.py | sha256=gGUnNbJTfTh5w6EDMr6VGE3H2iWyhqEYIEwt-iHB4a4 | 1404 |
llama_models/llama3/reference_impl/__init__.py | sha256=Pe9BV6DqLkla7ap6HtghozrqI1D-d3eW788jQQDGbbA | 276 |
llama_models/llama3/reference_impl/generation.py | sha256=cBr1R09jYQutkp81dnhkF6oTFuZy_erzhfOyBTYJZZ0 | 16683 |
llama_models/llama3/reference_impl/model.py | sha256=kR-wlvF1gON-awIp30tskxPHct3y9o4WqGRHrxewR0U | 11818 |
llama_models/llama3/reference_impl/multimodal/__init__.py | sha256=Pe9BV6DqLkla7ap6HtghozrqI1D-d3eW788jQQDGbbA | 276 |
llama_models/llama3/reference_impl/multimodal/encoder_utils.py | sha256=64XlHbD-mYNvbTwqAKqNeIBjjt1jWG-X7zTSx2KwXJY | 6474 |
llama_models/llama3/reference_impl/multimodal/image_transform.py | sha256=jo-bHzcfgchGYWY4PAncnBdJtDFSMZASl9UaT0ZBut4 | 16622 |
llama_models/llama3/reference_impl/multimodal/model.py | sha256=wqiwDnglf2wY7392XaQmqYl4xjyUth2Wd59_I6Qth94 | 52306 |
llama_models/llama3/reference_impl/multimodal/utils.py | sha256=--8CQZ2MEfYnka3SrL6kKk-YRgSMCVfIs6yevYEfZf0 | 484 |
llama_models/llama3/tests/api/test_generation.py | sha256=Dd05D7pFUhv2A7DoJpMK69fSmwVxTvgOmpf-R_YSqZM | 3489 |
llama_models/llama3/tests/api/test_tool_utils.py | sha256=bZvVDseuInl8-ZlDdQtlBzoCKfyKdYLGFlr5SOuES2Y | 5721 |
llama_models/llama3/tests/prompt_templates/test_system_prompts.py | sha256=cgT1hWMVqFz56Ct0KyqdBXGCXToKOqpcTcduwoLVrF0 | 6347 |
llama_models/llama3_1/__init__.py | sha256=Pe9BV6DqLkla7ap6HtghozrqI1D-d3eW788jQQDGbbA | 276 |
llama_models/llama3_1/prompt_format.md | sha256=f4Wi5ZmcD_gsqqHiYoq-vCTPhxuUVJdLRB3EF1bGFqQ | 11774 |
llama_models/llama3_1/prompts.py | sha256=TGQZnetERTlgZlg9DHAMOICWe8_uawsVPNyFcki75Wk | 11844 |
llama_models/llama3_2/__init__.py | sha256=Pe9BV6DqLkla7ap6HtghozrqI1D-d3eW788jQQDGbbA | 276 |
llama_models/llama3_2/prompts_text.py | sha256=sDKUh0pXIcSR3lXTpfDznezgdPeb0CR9IGLTfl7eOhk | 9370 |
llama_models/llama3_2/prompts_vision.py | sha256=c0MJZg8oCpLIKe_p6V6iXwsBZH_-nZQpKEzyBR3_nag | 5572 |
llama_models/llama3_2/text_prompt_format.md | sha256=M9MKMNuhbm56e65AxsVsG_JG0mUpDvqp3SCu41ApzUo | 9613 |
llama_models/llama3_2/vision_prompt_format.md | sha256=Xx42_esD3-fwfze8U2atkvYyLK1A9pvkJJlQjVIK3EI | 5047 |
llama_models/scripts/__init__.py | sha256=Pe9BV6DqLkla7ap6HtghozrqI1D-d3eW788jQQDGbbA | 276 |
llama_models/scripts/example_chat_completion.py | sha256=UsvwXINSgRhD_M8kFMb0ya0eyAQDkui7JgIyNAoRCCI | 3607 |
llama_models/scripts/example_text_completion.py | sha256=OjUCNlVono9Td3oZpC6RYZ-389Gslafh5pRMRN6eGCU | 1665 |
llama_models/scripts/generate_prompt_format.py | sha256=ftPIr_L8dHO4P2FG06jhMtiIixYkSXrD8E6lde4bIkY | 1524 |
llama_models/scripts/multimodal_example_chat_completion.py | sha256=9AYyXDtMeboMD62-Q1HQY8nLcFGJNhaeP4VM_EREGl4 | 2231 |
llama_models/scripts/multimodal_example_text_completion.py | sha256=jREhxKE9wV6Y0F6uSQEFQLX7NpY5O6GPPcZ2nut6DQo | 2019 |
llama_models/scripts/resources/dog.jpg | sha256=MOrU3JpVoSqMPipctb9ai3KFTbzd2rVoCSjVzEx0byY | 40215 |
llama_models/scripts/resources/pasta.jpeg | sha256=j7zVfjFge3ugK2Gi75wP6SCQCIeMOG5g2bt3VxfoEqs | 448611 |
llama_models-0.0.50.dist-info/LICENSE | sha256=ipqiNMdIuYMaIJgQliywQBIAokWKtC6LOAnZbTxxt8Y | 78 |
llama_models-0.0.50.dist-info/METADATA | sha256=PQHUo0z-Z5EEEghK5xgucQGriAo0uBRuI0Ut2kb7Ztc | 8251 |
llama_models-0.0.50.dist-info/WHEEL | sha256=P9jw-gEje8ByB7_hXoICnHtVCrEwMQh-630tKvQWehc | 91 |
llama_models-0.0.50.dist-info/entry_points.txt | sha256=3oPNIFyiF5DXZQ9YwhtfIzbzDXbcaV2OMXqD_XGuUO4 | 366 |
llama_models-0.0.50.dist-info/top_level.txt | sha256=1InuU2jdsy7Uo3a9GA6v8ljZOt8fF2gOfyeEq88wCxQ | 13 |
llama_models-0.0.50.dist-info/RECORD | — | — |
top_level.txt
llama_models
entry_points.txt
example_chat_completion = llama_models.scripts.example_chat_completion:main
example_text_completion = llama_models.scripts.example_text_completion:main
multimodal_example_chat_completion = llama_models.scripts.multimodal_example_chat_completion:main
multimodal_example_text_completion = llama_models.scripts.multimodal_example_text_completion:main