{"id":13393597,"url":"https://github.com/visual-openllm/visual-openllm","last_synced_at":"2025-05-16T04:05:58.745Z","repository":{"id":147938403,"uuid":"619028064","full_name":"visual-openllm/visual-openllm","owner":"visual-openllm","description":"something like visual-chatgpt, 文心一言的开源版","archived":false,"fork":false,"pushed_at":"2024-02-24T13:12:20.000Z","size":3089,"stargazers_count":1210,"open_issues_count":33,"forks_count":162,"subscribers_count":27,"default_branch":"main","last_synced_at":"2025-04-08T14:09:39.147Z","etag":null,"topics":[],"latest_commit_sha":null,"homepage":"","language":"Python","has_issues":true,"has_wiki":null,"has_pages":null,"mirror_url":null,"source_name":null,"license":null,"status":null,"scm":"git","pull_requests_enabled":true,"icon_url":"https://github.com/visual-openllm.png","metadata":{"files":{"readme":"README.md","changelog":null,"contributing":null,"funding":null,"license":null,"code_of_conduct":null,"threat_model":null,"audit":null,"citation":null,"codeowners":null,"security":null,"support":null,"governance":null,"roadmap":null,"authors":null,"dei":null,"publiccode":null,"codemeta":null}},"created_at":"2023-03-26T03:26:41.000Z","updated_at":"2025-03-31T13:12:09.000Z","dependencies_parsed_at":"2024-12-08T08:03:31.911Z","dependency_job_id":"6dd22c5f-2752-4e2a-9d64-3d2f9c64fceb","html_url":"https://github.com/visual-openllm/visual-openllm","commit_stats":null,"previous_names":[],"tags_count":0,"template":false,"template_full_name":null,"repository_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/visual-openllm%2Fvisual-openllm","tags_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/visual-openllm%2Fvisual-openllm/tags","releases_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/visual-openllm%2Fvisual-openllm/releases","manifests_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories/visual-openllm%2Fvisual-openllm/manifests","owner_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/owners/visual-openllm","download_url":"https://codeload.github.com/visual-openllm/visual-openllm/tar.gz/refs/heads/main","host":{"name":"GitHub","url":"https://github.com","kind":"github","repositories_count":254464895,"owners_count":22075570,"icon_url":"https://github.com/github.png","version":null,"created_at":"2022-05-30T11:31:42.601Z","updated_at":"2022-07-04T15:15:14.044Z","host_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub","repositories_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repositories","repository_names_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/repository_names","owners_url":"https://repos.ecosyste.ms/api/v1/hosts/GitHub/owners"}},"keywords":[],"created_at":"2024-07-30T17:00:56.671Z","updated_at":"2025-05-16T04:05:57.545Z","avatar_url":"https://github.com/visual-openllm.png","language":"Python","readme":"# Visual OpenLLM\n\n一种基于开源模型, 已交互方式连接不同视觉模型的开源工具。\n\n- 基于 ChatGLM + Visual ChatGPT + Stable Diffusion\n- 开源版的\"文心一言\"\n\n\n## 📝 Changelog\n\n- [2024.2.24] 增加对Chatglm3的支持，新增vqa和pix2pix功能。(thank [@MrChen314](https://github.com/MrChen314) for this contribution )\n- [2023.3.27] 代码开源\n\n\n## Demo：\n\n![](assets/demo.gif)\n\n\n## 运行:\n\nrun with `Chatglm3-6B`(Default)\n```\npython run.py --load_llm Chatglm3\n```\nrun with `Chatglm`\n\n```\npython run.py --load_llm Chatglm\n```\n\n## Todo:\n\n- 支持多轮chat\n- 支持其他视觉工具\n- 支持其他LLM\n","funding_links":[],"categories":["NLP","A01_文本生成_文本对话","精选开源项目合集","multi-modal","Python"],"sub_categories":["大语言对话模型及数据","GPT镜像平替","glm 6b"],"project_url":"https://awesome.ecosyste.ms/api/v1/projects/github.com%2Fvisual-openllm%2Fvisual-openllm","html_url":"https://awesome.ecosyste.ms/projects/github.com%2Fvisual-openllm%2Fvisual-openllm","lists_url":"https://awesome.ecosyste.ms/api/v1/projects/github.com%2Fvisual-openllm%2Fvisual-openllm/lists"}