Rocio Gonzalez-Diaz , Eduardo Paluzo-Hidalgo, Miguel A. Gutiérrez-Naranjo
{"title":"神经网络的代表性数据集","authors":"Rocio Gonzalez-Diaz , Eduardo Paluzo-Hidalgo, Miguel A. Gutiérrez-Naranjo","doi":"10.1016/j.endm.2018.06.016","DOIUrl":null,"url":null,"abstract":"<div><p>Neural networks present big popularity and success in many fields. The large training time process problem is a very important task nowadays. In this paper, a new approach to get over this issue based on reducing dataset size is proposed. Two algorithms covering two different <em>shape</em> notions are shown and experimental results are given.</p></div>","PeriodicalId":35408,"journal":{"name":"Electronic Notes in Discrete Mathematics","volume":"68 ","pages":"Pages 89-94"},"PeriodicalIF":0.0000,"publicationDate":"2018-07-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1016/j.endm.2018.06.016","citationCount":"5","resultStr":"{\"title\":\"Representative datasets for neural networks\",\"authors\":\"Rocio Gonzalez-Diaz , Eduardo Paluzo-Hidalgo, Miguel A. Gutiérrez-Naranjo\",\"doi\":\"10.1016/j.endm.2018.06.016\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><p>Neural networks present big popularity and success in many fields. The large training time process problem is a very important task nowadays. In this paper, a new approach to get over this issue based on reducing dataset size is proposed. Two algorithms covering two different <em>shape</em> notions are shown and experimental results are given.</p></div>\",\"PeriodicalId\":35408,\"journal\":{\"name\":\"Electronic Notes in Discrete Mathematics\",\"volume\":\"68 \",\"pages\":\"Pages 89-94\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2018-07-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://sci-hub-pdf.com/10.1016/j.endm.2018.06.016\",\"citationCount\":\"5\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Electronic Notes in Discrete Mathematics\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S1571065318301070\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"Mathematics\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Electronic Notes in Discrete Mathematics","FirstCategoryId":"1085","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1571065318301070","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"Mathematics","Score":null,"Total":0}
Neural networks present big popularity and success in many fields. The large training time process problem is a very important task nowadays. In this paper, a new approach to get over this issue based on reducing dataset size is proposed. Two algorithms covering two different shape notions are shown and experimental results are given.
期刊介绍:
Electronic Notes in Discrete Mathematics is a venue for the rapid electronic publication of the proceedings of conferences, of lecture notes, monographs and other similar material for which quick publication is appropriate. Organizers of conferences whose proceedings appear in Electronic Notes in Discrete Mathematics, and authors of other material appearing as a volume in the series are allowed to make hard copies of the relevant volume for limited distribution. For example, conference proceedings may be distributed to participants at the meeting, and lecture notes can be distributed to those taking a course based on the material in the volume.