Topic #10: 오픈소스 LLM 씬의 라이징 스타! 'DeepSeek'을 알아보자 > 자유게시판

Topic #10: 오픈소스 LLM 씬의 라이징 스타! 'DeepSeek'을 알아보자

페이지 정보

profile_image
작성자 Porfirio
댓글 0건 조회 9회 작성일 25-02-01 11:09

본문

The DeepSeek v3 paper (and are out, after yesterday's mysterious release of Loads of fascinating particulars in here. More evaluation outcomes could be found right here. That is probably solely mannequin specific, so future experimentation is required here. This mannequin is a effective-tuned 7B parameter LLM on the Intel Gaudi 2 processor from the Intel/neural-chat-7b-v3-1 on the meta-math/MetaMathQA dataset. The Intel/neural-chat-7b-v3-1 was originally high quality-tuned from mistralai/Mistral-7B-v-0.1. 1.3b-instruct is a 1.3B parameter model initialized from deepseek ai china-coder-1.3b-base and superb-tuned on 2B tokens of instruction information.

댓글목록

등록된 댓글이 없습니다.