• Repositorio Institucional Universidad de Pamplona
  • Producción Editorial Universidad de Pamplona
  • Revistas
  • Revista Colombiana de Tecnologias de Avanzada (RCTA)
  • Por favor, use este identificador para citar o enlazar este ítem: http://repositoriodspace.unipamplona.edu.co/jspui/handle/20.500.12744/10372
    Registro completo de metadatos
    Campo DC Valor Lengua/Idioma
    dc.contributor.authorMosquera de la Cruz, José Hernando-
    dc.contributor.authorLoaiza Correa, Humberto-
    dc.contributor.authorNope Rodríguez, Sandra Esperanza-
    dc.date.accessioned2025-10-14T20:54:41Z-
    dc.date.available2025-10-14T20:54:41Z-
    dc.date.issued2025-01-01-
    dc.identifier.citationJ. H. Mosquera Cruz, H. Loaiza Correa, y S. E. Nope Rodríguez, «Acceso a Internet por Comandos de Voz: Aplicación de Navegación para Facebook, Gmail y Chrome», RCTA, vol. 1, n.º 45, pp. 183–194, ene. 2025. https://doi.org/10.24054/rcta.v1i45.2963es_CO
    dc.identifier.issn1692-7257-
    dc.identifier.issn2500-8625-
    dc.identifier.urihttp://repositoriodspace.unipamplona.edu.co/jspui/handle/20.500.12744/10372-
    dc.descriptionSe presenta un sistema para navegar por Internet mediante comandos de voz. La herramienta implementada permitió el control verbal de las aplicaciones Google Chrome, Gmail y Facebook. Las pruebas se realizaron con un grupo de 33 personas con diferentes experiencias de navegación por Internet compuesto por adultos jóvenes, adultos mayores y personas con discapacidad motriz. Cada una de las aplicaciones se probó por separado mediante diálogos guiados con comandos de voz y dictados. En las pruebas del sistema de reconocimiento de voz se utilizaron 2871 comandos de voz y 594 dictados, observándose un mejor resultado de los comandos de voz en la aplicación Facebook y de los dictados en la aplicación Google Chrome. Se obtuvo una media general del 84,69% con una desviación estándar del 6,45% para el reconocimiento de comandos de voz, y del 74,63% con una desviación estándar del 2,75% para el reconocimiento de dictados.es_CO
    dc.description.abstractA system to navigate the internet using voice commands is presented. The implemented tool allowed verbal control of Google Chrome, Gmail and Facebook applications. The tests were conducted on a group of 33 people with different experiences browsing the Internet composed of young adults, older adults and people with motor disabilities. Each of the applications was tested separately using guided dialogues with voice commands and dictations. In the speech recognition system’s tests, 2871 voice commands and 594 dictations were used, observing a better result for voice commands in the Facebook application and dictation in the Google Chrome application. A general average of 84.69% with a standard deviation of 6.45% was obtained for the recognition of voice commands, and 74.63% with a standard deviation of 2.75% for the recognition of dictations.es_CO
    dc.format.extent12es_CO
    dc.format.mimetypeapplication/pdfes_CO
    dc.language.isoeses_CO
    dc.publisherAldo Pardo García, Revista Colombiana de Tecnologías de Avanzada, Universidad de Pamplona.es_CO
    dc.relation.ispartofseries183;194-
    dc.subjectinteracción humano-computadores_CO
    dc.subjecttecnologías de la información y la comunicaciónes_CO
    dc.subjectnavegación en internetes_CO
    dc.subjectreconocimiento del hablaes_CO
    dc.titleAcceso a Internet por Comandos de Voz: Aplicación de Navegación para Facebook, Gmail y Chromees_CO
    dc.title.alternativeInternet Access by Voice Commands: Navigation Application for Facebook, Gmail and Chromees_CO
    dc.typehttp://purl.org/coar/resource_type/c_2df8fbb1es_CO
    dc.description.editionVol. 1 Núm. 45 (2025): Enero – Junioes_CO
    dc.relation.referencesN. S. Sreekanth et al., “Multimodal interface for Effective Man Machine Interaction,” Media Convergence Handbook, Media Business and Innovation, vol. 2, pp. 249–264, 2016. doi: 10.1007/978-3-642-54487-3.es_CO
    dc.relation.referencesB. Basharirad and M. Moradhaseli, “Speech Emotion Recognition Methods: A Literature Review,” 2nd Int. Conf. Appl. Sci. Technol. 2017, 2017, doi: 10.1063/1.5005438.es_CO
    dc.relation.referencesH. Ibrahim and A. Varol, “A Study on Automatic Speech Recognition Systems,” 8th Int. Symp. Digit. Forensics Secur. ISDFS 2020, p. 1–5, https://doi.org/10.1109/ISDFS49300.2020.91162, 2020.es_CO
    dc.relation.referencesS. Raju, V. Jagtap, P. Kulkarni, M. Ravikanth, and M. Rafeeq, “Speech Recognition to Build Context: A Survey,” Int. Conf. Comput. Sci. Eng. Appl., p. 1–7, https://doi.org/10.1109/iccsea49143.2020.9132, 2020.es_CO
    dc.relation.referencesL. Oksana, T. Ihor, and L. Pavlo, “Navigation Assistive Application for the Visually Impaired People,” Proc. - 2020 IEEE 11th Int. Conf. Dependable Syst. Serv. Technol. DESSERT 2020, p. 320–325, https://doi.org/10.1109/DESSERT50317.2020, 2020.es_CO
    dc.relation.referencesL. Clark et al., “The State of Speech in HCI: Trends, Themes and Challenges,” Interact. Comput., vol. 31, no. 4, p. 349–371, https://doi.org/10.1093/iwc/iwz016, 2019.es_CO
    dc.relation.referencesApple, “Siri,” [Online - Accesed May 4, 2023]. [Online]. Available: http://www.apple.com/siri/es_CO
    dc.relation.referencesMicrosoft, “Cortana,” [Online - Accesed May 4, 2023]. [Online]. Available: https://www.microsoft.com/en-us/cortanaes_CO
    dc.relation.referencesGoogle, “Google Assistant,” [Online - Accesed May 4, 2023]. [Online]. Available: https://assistant.google.com/intl/es_es/es_CO
    dc.relation.referencesAmazon, “Alexa,” [Online - Accesed May 4, 2023]. [Online]. Available: https://developer.amazon.com/alexaes_CO
    dc.relation.referencesSamsung, “Bixby,” [Online - Accesed May 4, 2023]. [Online]. Available: https://www.samsung.com/us/explore/bixby/es_CO
    dc.relation.referencesCerence, “Cerence Drive,” [Online - Accesed May 4, 2023]. [Online]. Available: https://www.cerence.com/cerence-products/beyond-voicees_CO
    dc.relation.referencesAmazon, “Astro,” [Online - Accesed May 4, 2023]. [Online]. Available: https://www.amazon.com/-/es/Presentamos-Amazon-Astro/dp/B078NSDFSBes_CO
    dc.relation.referencesR. Sarikaya, “The Technology Behind Personal Digital Assistants,” IEEE Signal Process. Mag., vol. 34, p. 67–81, https://doi.org/10.1109/MSP.2016.2617341, 2017.es_CO
    dc.relation.referencesV. Kepuska and G. Bohouta, “Next-generation of virtual personal assistants (Microsoft Cortana, Apple Siri, Amazon Alexa and Google Home),” 2018 IEEE 8th Annu. Comput. Commun. Work. Conf. CCWC 2018, vol. 2018–Janua, no. c, pp. 99–103, 2018, doi: 10.1109/CCWC.2018.8301638.es_CO
    dc.relation.referencesE. Marvin, “Digital Assistant for the Visually Impaired,” 2020 Int. Conf. Artif. Intell. Inf. Commun. ICAIIC 2020, p. 723–728, https://doi.org/10.1109/ICAIIC48513.2020., 2020.es_CO
    dc.relation.referencesM. B. Chandu and K. Ganapathy, “Voice Controlled Human Assistence Robot,” 2020 6th Int. Conf. Adv. Comput. Commun. Syst. ICACCS 2020, p. 971–973, https://doi.org/10.1109/ICACCS48705.2020., 2020.es_CO
    dc.relation.referencesS. Faroom, M. Nauman, S. Yousaf, and S. Umer, “Literature Review on Home Automation System,” Int. Conf. Comput. Math. Eng. Technol., 2018, doi: 10.17148/IJARCCE.2017.63173.es_CO
    dc.relation.referencesP. Suesaowaluk, “Home Automation System Based Mobile Application,” 2nd World Symp. Artif. Intell., p. 97–102, https://doi.org/10.1109/wsai49636.2020.914, 2020.es_CO
    dc.relation.referencesN. H. Abdallah, E. Affes, Y. Bouslimani, M. Ghribi, A. Kaddouri, and M. Ghariani, “Smart Assistant Robot for Smart Home Management,” 1st Int. Conf. Commun. Control Syst. Signal Process., p. 317–321, https://doi.org/10.1109/ccssp49278.2020.9, 2020.es_CO
    dc.relation.referencesP. J. Rani, J. Bakthakumar, B. P. Kumaar, U. P. Kumaar, and S. Kumar, “Voice controlled home automation system using natural language processing (NLP) and internet of things (IoT),” ICONSTEM 2017 - Proc. 3rd IEEE Int. Conf. Sci. Technol. Eng. Manag., vol. 2018–Janua, pp. 368–373, 2018, doi: 10.1109/ICONSTEM.2017.8261311.es_CO
    dc.relation.referencesP. Dabre, R. Gonsalves, R. Chandvaniya, and A. V. Nimkar, “A Framework for System Interfacing of Voice User Interface for Personal Computers,” 3rd Int. Conf. Commun. Syst. Comput. IT Appl., p. 1–6, https://doi.org/10.1109/cscita47329.2020.9137, 2020.es_CO
    dc.relation.referencesV. Chayapathy, G. S. Anitha, and B. Sharath, “IOT based home automation by using personal assistant,” Proc. 2017 Int. Conf. Smart Technol. Smart Nation, SmartTechCon 2017, pp. 385–389, 2018, doi: 10.1109/SmartTechCon.2017.8358401.es_CO
    dc.relation.referencesL. P. De Oliveira, M. A. Wehrmeister, and A. S. De Oliveira, “Systematic Literature Review on Automotive Diagnostics,” Brazilian Symp. Comput. Syst. Eng. SBESC, vol. 2017–Novem, pp. 1–8, 2017, doi: 10.1109/SBESC.2017.7.es_CO
    dc.relation.referencesH. Zhang and C. Ye, “Human-Robot Interaction for Assisted Wayfinding of a Robotic Navigation Aid for the Blind,” 12th Int. Conf. Hum. Syst. Interact. (HSI), Richmond, VA, USA, p. 137–142, https://doi.org/10.1109/HSI47298.2019.894, 2019.es_CO
    dc.relation.referencesG. Lugano, “Virtual assistants and self-driving cars,” Proc. 2017 15th Int. Conf. ITS Telecommun. ITST 2017, pp. 1–5, 2017, doi: 10.1109/ITST.2017.7972192.es_CO
    dc.relation.referencesM. Kim, E. Seong, Y. Jwa, J. Lee, and S. Kim, “A Cascaded Multimodal Natural User Interface to Reduce Driver Distraction,” IEEE Access, vol. 8, p. 112969–112984, https://doi.org/10.1109/ACCESS.2020, 2020.es_CO
    dc.relation.referencesS. Estes, J. Helleberg, K. Long, M. Pollack, and M. Quezada, “Guidelines for speech interactions between pilot & cognitive assistant,” ICNS 2018 - Integr. Commun. Navig. Surveill. Conf., pp. 1–23, 2018, doi: 10.1109/ICNSURV.2018.8384965.es_CO
    dc.relation.referencesS. Nur, A. Mohamad, and K. Isa, “Assistive Robot for Speech Semantic Recognition System,” 2018 7th Int. Conf. Comput. Commun. Eng., p. 50–55, ISBN: 9781538669921, 2018.es_CO
    dc.relation.referencesM. A. Hossain, M. F. K. Khondakar, M. H. Sarowar, and M. J. U. Qureshi, “Design and Implementation of an Autonomous Wheelchair,” 2019 4th Int. Conf. Electr. Inf. Commun. Technol. EICT 2019, p. 1–5, https://doi.org/10.1109/EICT48899.2019.906885, 2019.es_CO
    dc.relation.referencesN. H. Khan, A. H. Arovi, H. Mahmud, K. Hasan, and H. A. Rubaiyeat, “Speech based text correction tool for the visually impaired,” pp. 150–155, ISBN: 978-1-4673-9930–2, 2015.es_CO
    dc.relation.referencesP. Bose, A. Malpthak, U. Bansal, and A. Harsola, “Digital assistant for the blind,” 2017 2nd Int. Conf. Converg. Technol. I2CT 2017, vol. 2017–Janua, no. 2015, pp. 1250–1253, 2017, doi: 10.1109/I2CT.2017.8226327.es_CO
    dc.relation.referencesK. Cofre, E. Molina, and G. Guerrero, “Voice controlled interface oriented memory loss assistance system for older adults,” Iber. Conf. Inf. Syst. Technol. Cist., p. 24–27, https://doi.org/10.23919/CISTI49556.2020.91, 2020.es_CO
    dc.relation.referencesJ.-H. Mosquera-DeLaCruz, S.-E. Nope-Rodríguez, A.-D. Restrepo-Girón, and H. Loaiza-Correa, “Internet Access by Voice Commands (Source Code),” [Online - Accesed Jun 17, 2023]. Accessed: Jun. 17, 2023. [Online]. Available: https://github.com/nandostiwar/Internet_Access_by_Voice_Commands.gites_CO
    dc.relation.referencesLogitech, “LogitechG430,” [Online - Accesed May 4, 2023]. [Online]. Available: https://www.logitechg.com/es-roam/products.html?searchclick=gaminges_CO
    dc.relation.referencesMicrosoft, “Bing Speech API,” [Online - Accesed May 4, 2023]. [Online]. Available: https://azure.microsoft.com/es-es/services/cognitive-services/speech-to-text/#overviewes_CO
    dc.relation.referencesM. Assefi, M. Wittie, and A. Knight, “Impact of Network Performance on Cloud Speech Recognition,” 2015, doi: 10.1109/ICCCN.2015.7288417.es_CO
    dc.relation.referencesA. Hannun et al., “Deep Speech: Scaling up end-to-end speech recognition,” Arxiv, pp. 1–12, 2014, doi: arXiv:1412.5567v2.es_CO
    dc.relation.referencesJ.-H. Mosquera-DeLaCruz, S.-E. Nope-Rodríguez, A.-D. Restrepo-Girón, and H. Loaiza-Correa, “Disability and Rehabilitation : Assistive Technology Human-computer multimodal interface to internet navigation,” Disabil. Rehabil. Assist. Technol., vol. 0, no. 0, p. 1–14, https://doi.org/10.1080/17483107.2020.179944, 2020.es_CO
    dc.relation.referencesC. AutoIt, “AutoIt Library,” [Online - Accesed May 4, 2023], 2023. [Online]. Available: https://www.autoitscript.com/site/autoit/es_CO
    dc.relation.referencesMicrosoft, “System Speech Synthesis,” [Online - Accesed May 4, 2023], 2023. [Online]. Available: https://msdn.microsoft.com/en-us/library/system.speech.synthesis.aspxes_CO
    dc.relation.referencesM. Canelli, D. Grasso, and M. King, “Methods and Metrics for the Evaluation of Dictation Systems: a Case Study,” Proc. Second Int. Conf. Lang. Resour. Eval., p. 1–7, SematicScholar Corpus ID: 15527622, 2000.es_CO
    dc.rights.accessrightshttp://purl.org/coar/access_right/c_abf2es_CO
    dc.type.coarversionhttp://purl.org/coar/resource_type/c_2df8fbb1es_CO
    Aparece en las colecciones: Revista Colombiana de Tecnologias de Avanzada (RCTA)

    Ficheros en este ítem:
    Fichero Descripción Tamaño Formato  
    Art18_V1_N45_2025_esp.pdfArt18_V1_N45_2025_esp576,35 kBAdobe PDFVisualizar/Abrir


    Los ítems de DSpace están protegidos por copyright, con todos los derechos reservados, a menos que se indique lo contrario.