Please use this identifier to cite or link to this item:
http://hdl.handle.net/10397/110994
| DC Field | Value | Language |
|---|---|---|
| dc.contributor | Department of Computing | - |
| dc.contributor | Mainland Development Office | - |
| dc.creator | Guo, S | - |
| dc.creator | Wang, H | - |
| dc.creator | Zhan, Y | - |
| dc.date.accessioned | 2025-02-17T01:35:00Z | - |
| dc.date.available | 2025-02-17T01:35:00Z | - |
| dc.identifier.uri | http://hdl.handle.net/10397/110994 | - |
| dc.language.iso | zh | en_US |
| dc.publisher | 中华人民共和国国家知识产权局 | en_US |
| dc.rights | Assignee: 香港理工大学深圳研究院 | en_US |
| dc.title | Distributed deep learning method and device, terminal equipment and storage medium | en_US |
| dc.type | Patent | en_US |
| dc.description.otherinformation | Inventor name used in this publication: 郭嵩 | en_US |
| dc.description.otherinformation | Inventor name used in this publication: 王号召 | en_US |
| dc.description.otherinformation | Inventor name used in this publication: 詹玉峰 | en_US |
| dc.description.otherinformation | Title in Traditional Chinese: 一種分布式深度學習方法、裝置、終端設備及存儲介質 | en_US |
| dcterms.abstract | The invention is suitable for the technical field of computers, and provides a distributed deep learning method and device, terminal equipment and a storage medium, and the method comprises the steps: obtaining at least two same sample sets, each sample set comprising a plurality of data blocks; associating each data block with a working node according to a preset rule to ensure that the data blocks associated with any preset target number of working nodes can form at least one sample set, performing model training on the basis of each associated data block by the working node to obtain a gradient corresponding to the data block, and sending the gradient to a parameter server; the parameter server receives the gradients corresponding to the data blocks sent by the working node, calculates a target gradient based on the received gradients after receiving the gradients corresponding to all the data blocks in the at least one sample set, and sends the target gradient to the working node; the method does not affect the model training, improves the training speed, guarantees the integrity of the model training, and improves the accuracy of the model. | - |
| dcterms.abstract | 本申请适用于计算机技术领域,提供了一种分布式深度学习方法、装置、终端设备及存储介质,该方法包括:获取相同的至少两个样本集,每个样本集中包括多个的数据块;将各个数据块按照预设规则与工作节点关联,以保证任意预设目标数量的工作节点关联的数据块能够组成至少一个所述样本集,工作节点基于关联的各个数据块进行模型训练得到数据块对应的梯度,并向参数服务器发送所述梯度;参数服务器接收工作节点发送的所述数据块对应的梯度,并在接收到至少一个样本集中所有数据块对应的梯度后,基于接收到的梯度计算目标梯度,并向工作节点发送目标梯度;本申请不会影响模型训练,提高了训练速度,保证了模型训练的完整性,进而提高了模型的准确度。 | - |
| dcterms.accessRights | open access | en_US |
| dcterms.alternative | 一种分布式深度学习方法、装置、终端设备及存储介质 | - |
| dcterms.bibliographicCitation | 中国专利 ZL 202011018776.8 | - |
| dcterms.issued | 2024-08-06 | - |
| dc.description.country | China | - |
| dc.description.validate | 202502 bcch | - |
| dc.description.oa | Version of Record | en_US |
| dc.description.pubStatus | Published | en_US |
| dc.description.oaCategory | NA | en_US |
| Appears in Collections: | Patent | |
Files in This Item:
| File | Description | Size | Format | |
|---|---|---|---|---|
| ZL202011018776.8.pdf | 1.42 MB | Adobe PDF | View/Open |
Page views
4
Citations as of Apr 14, 2025
Downloads
51
Citations as of Apr 14, 2025
Google ScholarTM
Check
Items in DSpace are protected by copyright, with all rights reserved, unless otherwise indicated.


