Please use this identifier to cite or link to this item: http://hdl.handle.net/10397/110994
PIRA download icon_1.1View/Download Full Text
DC FieldValueLanguage
dc.contributorDepartment of Computing-
dc.contributorMainland Development Office-
dc.creatorGuo, S-
dc.creatorWang, H-
dc.creatorZhan, Y-
dc.date.accessioned2025-02-17T01:35:00Z-
dc.date.available2025-02-17T01:35:00Z-
dc.identifier.urihttp://hdl.handle.net/10397/110994-
dc.language.isozhen_US
dc.publisher中华人民共和国国家知识产权局en_US
dc.rightsAssignee: 香港理工大学深圳研究院en_US
dc.titleDistributed deep learning method and device, terminal equipment and storage mediumen_US
dc.typePatenten_US
dc.description.otherinformationInventor name used in this publication: 郭嵩en_US
dc.description.otherinformationInventor name used in this publication: 王号召en_US
dc.description.otherinformationInventor name used in this publication: 詹玉峰en_US
dc.description.otherinformationTitle in Traditional Chinese: 一種分布式深度學習方法、裝置、終端設備及存儲介質en_US
dcterms.abstractThe invention is suitable for the technical field of computers, and provides a distributed deep learning method and device, terminal equipment and a storage medium, and the method comprises the steps: obtaining at least two same sample sets, each sample set comprising a plurality of data blocks; associating each data block with a working node according to a preset rule to ensure that the data blocks associated with any preset target number of working nodes can form at least one sample set, performing model training on the basis of each associated data block by the working node to obtain a gradient corresponding to the data block, and sending the gradient to a parameter server; the parameter server receives the gradients corresponding to the data blocks sent by the working node, calculates a target gradient based on the received gradients after receiving the gradients corresponding to all the data blocks in the at least one sample set, and sends the target gradient to the working node; the method does not affect the model training, improves the training speed, guarantees the integrity of the model training, and improves the accuracy of the model.-
dcterms.abstract本申请适用于计算机技术领域,提供了一种分布式深度学习方法、装置、终端设备及存储介质,该方法包括:获取相同的至少两个样本集,每个样本集中包括多个的数据块;将各个数据块按照预设规则与工作节点关联,以保证任意预设目标数量的工作节点关联的数据块能够组成至少一个所述样本集,工作节点基于关联的各个数据块进行模型训练得到数据块对应的梯度,并向参数服务器发送所述梯度;参数服务器接收工作节点发送的所述数据块对应的梯度,并在接收到至少一个样本集中所有数据块对应的梯度后,基于接收到的梯度计算目标梯度,并向工作节点发送目标梯度;本申请不会影响模型训练,提高了训练速度,保证了模型训练的完整性,进而提高了模型的准确度。-
dcterms.accessRightsopen accessen_US
dcterms.alternative一种分布式深度学习方法、装置、终端设备及存储介质-
dcterms.bibliographicCitation中国专利 ZL 202011018776.8-
dcterms.issued2024-08-06-
dc.description.countryChina-
dc.description.validate202502 bcch-
dc.description.oaVersion of Recorden_US
dc.description.pubStatusPublisheden_US
dc.description.oaCategoryNAen_US
Appears in Collections:Patent
Files in This Item:
File Description SizeFormat 
ZL202011018776.8.pdf1.42 MBAdobe PDFView/Open
Open Access Information
Status open access
File Version Version of Record
Show simple item record

Page views

4
Citations as of Apr 14, 2025

Downloads

51
Citations as of Apr 14, 2025

Google ScholarTM

Check


Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.