{"status":"ok","message-type":"work","message-version":"1.0.0","message":{"indexed":{"date-parts":[[2024,9,13]],"date-time":"2024-09-13T00:26:39Z","timestamp":1726187199610},"publisher-location":"Cham","reference-count":19,"publisher":"Springer Nature Switzerland","isbn-type":[{"type":"print","value":"9783031258244"},{"type":"electronic","value":"9783031258251"}],"license":[{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"tdm","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"},{"start":{"date-parts":[[2023,1,1]],"date-time":"2023-01-01T00:00:00Z","timestamp":1672531200000},"content-version":"vor","delay-in-days":0,"URL":"https:\/\/www.springernature.com\/gp\/researchers\/text-and-data-mining"}],"content-domain":{"domain":["link.springer.com"],"crossmark-restriction":false},"short-container-title":[],"published-print":{"date-parts":[[2023]]},"DOI":"10.1007\/978-3-031-25825-1_13","type":"book-chapter","created":{"date-parts":[[2023,2,3]],"date-time":"2023-02-03T19:02:52Z","timestamp":1675450972000},"page":"175-187","update-policy":"http:\/\/dx.doi.org\/10.1007\/springer_crossmark_policy","source":"Crossref","is-referenced-by-count":0,"title":["Assessing the\u00a0Condition of\u00a0Copper Conductors Using Deep Learning"],"prefix":"10.1007","author":[{"ORCID":"http:\/\/orcid.org\/0000-0003-0712-9395","authenticated-orcid":false,"given":"Zhicheng","family":"Pan","sequence":"first","affiliation":[]},{"ORCID":"http:\/\/orcid.org\/0000-0003-0802-1845","authenticated-orcid":false,"given":"David","family":"Wilson","sequence":"additional","affiliation":[]},{"ORCID":"http:\/\/orcid.org\/0000-0002-7297-0351","authenticated-orcid":false,"given":"Martin","family":"Stommel","sequence":"additional","affiliation":[]},{"given":"Alex","family":"Castellanos","sequence":"additional","affiliation":[]}],"member":"297","published-online":{"date-parts":[[2023,2,4]]},"reference":[{"key":"13_CR1","unstructured":"EPRI. Parameters that influence the aging and degradation of overhead conductors. Technical report, EPRI (2003)"},{"key":"13_CR2","unstructured":"Graham, M., et al.: Distribution overhead copper conductors, their condition and risk-based replacement. Electricity Engineers\u2019 Association (2021)"},{"key":"13_CR3","doi-asserted-by":"publisher","first-page":"218863","DOI":"10.1109\/ACCESS.2020.3042486","volume":"8","author":"L Naranpanawe","year":"2020","unstructured":"Naranpanawe, L., et al.: A practical health index for overhead conductors: experience from Australian distribution networks. IEEE Access 8, 218863\u2013218873 (2020). https:\/\/doi.org\/10.1109\/ACCESS.2020.3042486. ISSN 2169-3536","journal-title":"IEEE Access"},{"key":"13_CR4","unstructured":"Van der Maaten, L., Hinton, G.: Visualizing data using t-SNE. J. Mach. Learn. Res. 9(11) (2008)"},{"key":"13_CR5","doi-asserted-by":"crossref","unstructured":"Selvaraju, R.R., et al.: Grad-CAM: visual explanations from deep networks via gradient-based localization. In: Proceedings of the IEEE International Conference on Computer Vision, pp. 618\u2013626 (2017)","DOI":"10.1109\/ICCV.2017.74"},{"key":"13_CR6","doi-asserted-by":"publisher","unstructured":"Ishino, R., Tsutsumi, F.: Detection system of damaged cables using video obtained from an aerial inspection of transmission lines. In: IEEE Power Engineering Society General Meeting, vol. 2, pp. 1857\u20131862 (2004). https:\/\/doi.org\/10.1109\/PES.2004.1373201","DOI":"10.1109\/PES.2004.1373201"},{"key":"13_CR7","doi-asserted-by":"publisher","first-page":"59022","DOI":"10.1109\/ACCESS.2019.2914766","volume":"7","author":"Y Zhang","year":"2019","unstructured":"Zhang, Y., et al.: A recognition technology of transmission lines conductor break and surface damage based on aerial image. IEEE Access 7, 59022\u201359036 (2019). https:\/\/doi.org\/10.1109\/ACCESS.2019.2914766. ISSN 2169-3536","journal-title":"IEEE Access"},{"issue":"11","key":"13_CR8","doi-asserted-by":"publisher","first-page":"8783","DOI":"10.1109\/TIM.2020.2994475","volume":"69","author":"X Huang","year":"2020","unstructured":"Huang, X., et al.: A method of transmission conductor-loosened detect based on image sensors. IEEE Trans. Instrum. Meas. 69(11), 8783\u20138796 (2020). https:\/\/doi.org\/10.1109\/TIM.2020.2994475","journal-title":"IEEE Trans. Instrum. Meas."},{"issue":"5","key":"13_CR9","doi-asserted-by":"publisher","first-page":"2154","DOI":"10.1109\/TPWRD.2014.2328572","volume":"29","author":"Y Song","year":"2014","unstructured":"Song, Y., Wang, H., Zhang, J.: A vision-based broken strand detection method for a power-line maintenance robot. IEEE Trans. Power Deliv. 29(5), 2154\u20132161 (2014). https:\/\/doi.org\/10.1109\/TPWRD.2014.2328572. ISSN 1937-4208","journal-title":"IEEE Trans. Power Deliv."},{"key":"13_CR10","unstructured":"Glorot, X., Bengio, Y.: Understanding the difficulty of training deep feedforward neural networks. In: AISTATS. JMLR Proceedings, vol. 9, pp. 249\u2013256. JMLR.org (2010)"},{"key":"13_CR11","unstructured":"Goodfellow, I., Bengio, Y., Courville, A.: Deep Learning. MIT Press, Cambridge (2016). https:\/\/www.deeplearningbook.org"},{"key":"13_CR12","unstructured":"M\u00fcller, R., Kornblith, S., Hinton, G.E.: When does label smoothing help? In: Advances in Neural Information Processing Systems, vol. 32 (2019)"},{"key":"13_CR13","unstructured":"Krizhevsky, A., Sutskever, I., Hinton, G.E.: Imagenet classification with deep convolutional neural networks. In: Proceedings of the 25th International Conference on Neural Information Processing Systems - Volume 1, NIPS 2012, Lake Tahoe, Nevada, pp. 1097\u20131105. Curran Associates Inc. (2012)"},{"key":"13_CR14","doi-asserted-by":"publisher","unstructured":"Iandola, F.N., et al.: Squeezenet: Alexnet-level accuracy with 50x fewer parameters and $$< 0.5$$ MB model size (2016). https:\/\/doi.org\/10.48550\/ARXIV.1602.07360","DOI":"10.48550\/ARXIV.1602.07360"},{"key":"13_CR15","doi-asserted-by":"publisher","unstructured":"Szegedy, C., et al.: Going deeper with convolutions. In: 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1\u20139 (2015). https:\/\/doi.org\/10.1109\/CVPR.2015.7298594","DOI":"10.1109\/CVPR.2015.7298594"},{"key":"13_CR16","doi-asserted-by":"publisher","unstructured":"He, K., et al.: Deep residual learning for image recognition. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 770\u2013778 (2016). https:\/\/doi.org\/10.1109\/CVPR.2016.90","DOI":"10.1109\/CVPR.2016.90"},{"key":"13_CR17","doi-asserted-by":"crossref","unstructured":"Sandler, M., et al.: Mobilenetv 2: inverted residuals and linear bottlenecks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 4510\u20134520 (2018)","DOI":"10.1109\/CVPR.2018.00474"},{"key":"13_CR18","doi-asserted-by":"crossref","unstructured":"Szegedy, C., et al.: Inception-v4, inception-resnet and the impact of residual connections on learning. In: Thirty-First AAAI Conference on Artificial Intelligence (2017)","DOI":"10.1609\/aaai.v31i1.11231"},{"key":"13_CR19","unstructured":"Springenberg, J.T., et al.: Striving for simplicity: the all convolutional net. In: Bengio, Y., LeCun, Y. (eds.) 3rd International Conference on Learning Representations, ICLR 2015, San Diego, CA, USA, 7\u20139 May 2015, Workshop Track Proceedings (2015). https:\/\/arxiv.org\/abs\/1412.6806"}],"container-title":["Lecture Notes in Computer Science","Image and Vision Computing"],"original-title":[],"language":"en","link":[{"URL":"https:\/\/link.springer.com\/content\/pdf\/10.1007\/978-3-031-25825-1_13","content-type":"unspecified","content-version":"vor","intended-application":"similarity-checking"}],"deposited":{"date-parts":[[2023,2,3]],"date-time":"2023-02-03T19:06:01Z","timestamp":1675451161000},"score":1,"resource":{"primary":{"URL":"https:\/\/link.springer.com\/10.1007\/978-3-031-25825-1_13"}},"subtitle":[],"short-title":[],"issued":{"date-parts":[[2023]]},"ISBN":["9783031258244","9783031258251"],"references-count":19,"URL":"https:\/\/doi.org\/10.1007\/978-3-031-25825-1_13","relation":{},"ISSN":["0302-9743","1611-3349"],"issn-type":[{"type":"print","value":"0302-9743"},{"type":"electronic","value":"1611-3349"}],"subject":[],"published":{"date-parts":[[2023]]},"assertion":[{"value":"4 February 2023","order":1,"name":"first_online","label":"First Online","group":{"name":"ChapterHistory","label":"Chapter History"}},{"value":"IVCNZ","order":1,"name":"conference_acronym","label":"Conference Acronym","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"International Conference on Image and Vision Computing New Zealand","order":2,"name":"conference_name","label":"Conference Name","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Auckland","order":3,"name":"conference_city","label":"Conference City","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"New Zealand","order":4,"name":"conference_country","label":"Conference Country","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"2022","order":5,"name":"conference_year","label":"Conference Year","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"23 November 2022","order":7,"name":"conference_start_date","label":"Conference Start Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"24 November 2022","order":8,"name":"conference_end_date","label":"Conference End Date","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"37","order":9,"name":"conference_number","label":"Conference Number","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"ivcnz2022","order":10,"name":"conference_id","label":"Conference ID","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"https:\/\/ivcnz2022.aut.ac.nz\/","order":11,"name":"conference_url","label":"Conference URL","group":{"name":"ConferenceInfo","label":"Conference Information"}},{"value":"Double-blind","order":1,"name":"type","label":"Type","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"EasyChair","order":2,"name":"conference_management_system","label":"Conference Management System","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"79","order":3,"name":"number_of_submissions_sent_for_review","label":"Number of Submissions Sent for Review","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"14","order":4,"name":"number_of_full_papers_accepted","label":"Number of Full Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"23","order":5,"name":"number_of_short_papers_accepted","label":"Number of Short Papers Accepted","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"18% - The value is computed by the equation \"Number of Full Papers Accepted \/ Number of Submissions Sent for Review * 100\" and then rounded to a whole number.","order":6,"name":"acceptance_rate_of_full_papers","label":"Acceptance Rate of Full Papers","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"2.7","order":7,"name":"average_number_of_reviews_per_paper","label":"Average Number of Reviews per Paper","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"3.1","order":8,"name":"average_number_of_papers_per_reviewer","label":"Average Number of Papers per Reviewer","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}},{"value":"No","order":9,"name":"external_reviewers_involved","label":"External Reviewers Involved","group":{"name":"ConfEventPeerReviewInformation","label":"Peer Review Information (provided by the conference organizers)"}}]}}