全國中小學科展

電腦科學與資訊工程

Method of prosthetic vision

This work is devoted to solving the problem of orientation in the space of visually impaired people. Working on the project, a new way of transmitting visual information through an acoustic channel was invented. In addition, was developed the device, which uses distance sensors to analyze the situation around a user. Thanks to the invented algorithm of transformation of the information about the position of the obstacle into the sound of a certain tone and intensity, this device allows the user to transmit subject-spatial information in real time. Currently, the device should use a facette locator made of 36 ultrasonic locators grouped in 12 sectors by the azimuth and 3 spatial cones by the angle. Data obtained in such a way is converted into its own note according to the following pattern : the angle of the place corresponds to octave, the azimuth corresponds to the note and the distance corresponds to the volume. The choice of the notes is not unambiguous. However, we used them for the reason that over the centuries, notes have had a felicitous way of layout on the frequency range and on the logarithmic scale. Therefore, the appearance of a new note in the total signal will not be muffled by a combination of other notes. Consequently, a blind person, moving around the room with the help of the tone and volume of the sound signals, will be able to assess the presence and location of all dangerous obstacles. After theoretical substantiation of the hypothesis and analysis of the available information, we started the production of prototypes of the devices that would implement the idea of transmitting information via the acoustic channel.

第五代行動通訊中基地台毫米波天線精確的方位角量測

第五代行動通訊(5th generation mobile networks)是現今科技發展的趨勢,新技術的出現也衍生出很多新的問題,在基地台點對點傳輸時,需要精確角度的天線才足以準確地接收高頻波短的毫米波,雖然現今已經有精密儀器能測量精確的方位角,但價格較高且使用方法複雜,面對數量龐大的5G基地台時,維修成本過高。本研究利用手機拍照得到天線與目標物相對角度,結合預先得知目標物的方位角,再經過數學運算即可得到精確的天線指向。本研究希望以隨手可得的手機,配合簡單的方法,可得到精確的天線指向,解決第五代行動通訊可能面臨的問題。

Development of an Android Application for Triage Prediction in Hospital Emergency Departments

Triage is the process by which nurses manage hospital emergency departments by assigning patients varying degrees of urgency. While triage algorithms such as the Emergency Severity Index (ESI) have been standardized worldwide, many of them are highly inconsistent, which could endanger the lives of thousands of patients. One way to improve on nurses’ accuracy is to use machine learning models (ML), which can learn from past data to make predictions. We tested six ML models: random forest, XGBoost, logistic regression, support vector machines, k-nearest neighbors, and multilayer perceptron. These models were tasked with predicting whether a patient would be admitted to the intensive care unit (ICU), another unit in the hospital, or be discharged. After training on data from more than 30,000 patients and testing using 10-fold cross-validation, we found that all six models outperformed ESI. Of the six, the random forest model achieved the highest average accuracy in predicting both ICU admission (81% vs. 69% using ESI; p<0.001) and hospitalization (75% vs. 57%; p<0.001). These models were then added to an Android application, which would accept patient data, predict their triage, and then add them to a priority-ordered waiting list. This approach may offer significant advantages over conventional triage: mainly, it has a higher accuracy than nurses and returns predictions instantaneously. It could also stand-in for triage nurses entirely in disasters, where medical personnel must deal with a large influx of patients in a short amount of time.

利用Yolo 模型辨識台灣國語口手語之研究

手語為聾啞人士日常溝通的工具,但對一般人來說這是一種難以理解的溝通方式。本實驗使用深度學習的 Yolov3 與 Yolov4 模型訓練37個國語注音符號手勢,然後再驗證模型對圖片、影片、即時(Real time)攝影辨識的正確率。 實驗結果顯示:Yolo v3 圖片辨識度效果還不錯,但影片辨識度很差,而Yolo v4 不管在靜態的圖片或動態影片都有不錯的辨識率,另外在即時的影像辨識也有不錯的效果。 雖然有部分符號的辨識度很低,但這可能是訓練時照片拍攝的問題,如果可以改進拍攝的數量和技巧,相信可以大幅提升判讀的準確率。

Cross-lingual Information Retrieval

In this project, we evaluate the effectiveness of Random Shuffling in the Cross Lingual Information Retrieval (CLIR) process. We extended the monolingual Word2Vec model to a multilingual one via the random shuffling process. We then evaluate the cross-lingual word embeddings (CLE) in terms of retrieving parallel sentences, whereby the query sentence is in a source language and the parallel sentence is in some targeted language. Our experiments on three language pairs showed that models trained on a randomly shuffled dataset outperforms randomly initialized word embeddings substantially despite its simplicity. We also explored Smart Shuffling, a more sophisticated CLIR technique which makes use of word alignment and bilingual dictionaries to guide the shuffling process, making preliminary comparisons between the two. Due to the complexity of the implementation and unavailability of open source codes, we defer experimental comparisons to future work.

彩色二維條碼手持產品開發之探討

QR Code是由黑白模組組成的二維數位條碼,掃描後可讀取儲存的訊息。受限於設計原理,QR Code使用二進位制儲存資料。增加模組數目可增加資料量,但若在條碼內塞進太多模組時,尺寸太小的模組將無法被掃描器讀取。此外,目前QR Code掃描器僅支援單張掃描,並無法應付同時多張條碼掃描的實務需求。 如能克服顏色辨識,理論上彩色二維條碼將能克服現行QR Code的限制,但市面上並無相關產品可供測試。因此本專題設計了一款10×10、具8顏色的"Colour Matrix",並利用Raspberry Pi開發Colour Matrix在手持裝置上運作的軟硬體來進行實驗。此實驗成功利用機器學習演算法在Raspberry Pi上進行的顏色辨識。開發的程式在單張掃描上效能與使用pyzbar辨識QR Code相當;在多張掃描方面,使用pyzbar辨識QR Code的解碼成功率為3.1%,而本專題的方法將成功率提升至92.4%,擴增數位條碼的使用範圍,具商用價值。

語音情緒辨識之研究

情緒辨識是增進人際溝通的重要能力。如生命線、電話客服等應用情境缺乏表情、肢體語言等輔助時,單以語音進行情緒辨識有極高的實用價值。 本研究探討比較支持向量機(SVM)及卷積神經網路(CNN)兩種機器學習方法於訓練「AI語音情緒辨識」分類器模型的表現。我們採用SAVEE和RAVDESS兩個英文語音資料庫,並自行製作與標註「逼逼中文情緒語料庫」。研究結果顯示SVM對SAVEE資料庫單一情緒的辨識正確率達84~94%,個別錄音員正確率達75%,超越官網紀錄的73.7%。同時,實驗顯示深度學習的模型在訓練資料不足的狀況下,反而相對遜色。

Development of an autonomous Search and Rescue Drone

The number of natural disasters has risen significantly in recent years, and with climate change there is no end in sight. Consequently, the demands on rescue forces around the world are increasing. For this reason, I asked myself what I can do to improve the work of rescue teams. Advances in artificial intelligence and drone technology enable new possibilities for problem solving. Based on the technological advances mentioned above, an autonomous Search and Rescue drone was developed as part of this project. The system assists rescue workers in searching for survivors of natural disasters or missing people. This paper also suggests a method for prioritizing survivors based on their vitality. The system was implemented using a commercial Parrot ANAFI drone and Python. The software was tested on a simulated drone. To simplify the development, the whole system was divided into the following subsystems: Navigation System, Search System and Mission Abort System. These subsystems were tested independently. The testing of solutions and new concepts were performed using smaller test programs on the simulated drone and finally on the physical drone. The Search and Rescue system was successfully developed. The person detection system can detect humans and distinguish them from the environment. Furthermore, based on the movements of a person, the system can distinguish whether the person is a rescuer or a victim. In addition, an area to be flown over can be defined. If something goes wrong during the mission, the mission can be aborted by the Mission Abort System. In the simulation, the predefined area can successfully be flown over. Unfortunately, controlling the physical drone does not work. It stops in the air after takeoff due to the firmware of the drone. It does not change the flight state of the drone, which results in all subsequent commands from the system being ignored. This paper shows that artificial intelligence and drone technologies can be combined to deliver better rescue services. The same system can be applied to other applications.

Cross-lingual Information Retrieval

In this project, we evaluate the effectiveness of Random Shuffling in the Cross Lingual Information Retrieval (CLIR) process. We extended the monolingual Word2Vec model to a multilingual one via the random shuffling process. We then evaluate the cross-lingual word embeddings (CLE) in terms of retrieving parallel sentences, whereby the query sentence is in a source language and the parallel sentence is in some targeted language. Our experiments on three language pairs showed that models trained on a randomly shuffled dataset outperforms randomly initialized word embeddings substantially despite its simplicity. We also explored Smart Shuffling, a more sophisticated CLIR technique which makes use of word alignment and bilingual dictionaries to guide the shuffling process, making preliminary comparisons between the two. Due to the complexity of the implementation and unavailability of open source codes, we defer experimental comparisons to future work.

以深度學習與遷移學習防範社群媒體片面新聞訊息之研究

現代民眾獲取新聞的途徑逐漸轉移到網路媒體,然而在網路資訊快速傳播以及媒體為追求報導曝光度以增加金錢利益的情形下,片面、誘導等形式的新聞標題與短句訊息在新聞媒體傳播中日益嚴重;本次研究透過Fake News Challenge提供的Stance Detection dataset,運用深度學習與遷移學習方法訓練可預測兩文本之間相關程度的自然語言處理模型,在過程中改善調參及訓練方式,並將其實際運用在預測美國新聞媒體於Facebook網路社群平台發文推播新聞的同時所附的短句與新聞報導文本內容之間的相關關係程度,分析社群平台中新聞可能造成的誤導式文句是否實際造成片面報導,而影響了受眾對於媒體的使用程度與信任程度。使此模型有助即時預警社群平台上的報導資訊型態品質,輔助使用者獲取新聞時所應具備的媒體識讀能力,進而改善片面報導於網路的流竄,同時提升未來媒體生態。