Speech recognition is of an important contribution in promoting new technologies in human computer interaction. Today, there is a growing need to employ speech technology in daily life and business activities. However, speech recognition is a challenging task that requires different stages before obtaining the desired output. Among automatic speech recognition (ASR) components is the feature extraction process, which parameterizes the speech signal to produce the corresponding feature vectors. Feature extraction process aims at approximating the linguistic content that is conveyed by the input speech signal. In speech processing field, there are several methods to extract speech features, however, Mel Frequency Cepstral Coefficients (MFCC) is the popular technique. It has been long observed that the MFCC is dominantly used in the well-known recognizers such as the Carnegie Mellon University (CMU) Sphinx and the Markov Model Toolkit (HTK). Hence, this paper focuses on the MFCC method as the standard choice to identify the different speech segments in order to obtain the language phonemes for further training and decoding steps. Due to MFCC good performance, the previous studies show that the MFCC dominates the Arabic ASR research. In this paper, we demonstrate MFCC as well as the intermediate steps that are performed to get these coefficients using the HTK toolkit.
For readers, space/place is one of the most significant references to reveal deep significances and indications in modern Arabic poetic texts. Generally, when poets evoke places and/or spaces, they do not mean to refer readers to detailed geographic or physical spaces, but to the symbolic significances and dimensions that those spaces have and through which poets encourage spacial awareness in their readers. Recently, as a result, there has been a great deal of interest in research addressing spacial poetic texts and dimensions in modern Arabic poetry in general and in Palestinian poetry in particular. Samih al-Qasim is one of the most recent prominent Palestinian revolutionary poets. Al-Qasim has published six series of poems that are well known in the Arab world. Although several researchers have studied al-Qasim's poetry, to our knowledge, yet no one has studied the aspect of spacial poetic text in his poetry. Therefore, this paper seeks to fill a gap in the scholarship that has not been addressed up to now. This article aims, not only to demonstrate the presence of spacial poetic text and dimensions throughout al-Qasim's poetry, but also to investigate the purpose for which the poet uses spacial poetic text. Our theory is that the poet, consciously and significantly, uses spacial poetic texts to magnify the Palestinian identity of the Palestinian readers. Methodologically, we applied a descriptive analytic method, referencing al-Qasim's poetry, addressing spacial poetic texts practically but not theoretically or statistically.
An analysis of the five most used numerals and a proposal for the adoption of a Universally Acceptable Numerals (UAN), came up as a result of the researchers inquisitiveses of the need for a set of numerals that is universally accepted. The researcher sought for the meaning of the first letter, “Nun”, “ن”, of the first verse of Suratul-Kalam (Chapter of the Pen), the Sixty-Eighth Chapter of the Holy Qur'an. It was observed that there was no universally accepted, economical, explainable, linkable and consistent set of numerals used by all scientists up till the moment of making this enquiry. As a theoretical paper, explanatory method is used to review five of the most used numerals (Tally Marks, Roman Figure, Hindu-Arabic, Arabic, and Chinese) and the urgent need for a universally accepted, economical, explainable, linkable and consistent set of numerals arises. The study discovers: ., I, \, _, L, U, =, C, O, 9, and 1.; to be used as numeral 0, 1, 2, 3, 4, 5, 6, 7, 8, 9 and 10 respectively; as a set of universally acceptable, economical, explainable, linkable, sustainable, convertible and consistent set of numerals that originates from Islam. They can be called Islameconumerals or UAN. With UAN, everything dropped, written, drawn and/or scribbled has meaning(s) as postulated by the first verse of Qur'an 68 and everyone can easily document all figures within the shortest period. It is suggested that there should be a discipline called Numeralnomics (Study of optimum utilization of Numerals) and everybody should start using the UAN, now, in order in know their strengths and weaknesses so as to suggest a better and acceptable set of numerals for the interested readers. Similarly study can be conducted for the alphabets.
Arabs have a rhetorical heritage which has greatly contributed to the monitoring and analyzing of the rhetoric of the Holy Quran, Hadith, and Arabic texts on poetry and oratory. But Arab scholars - as far as the researcher knows – have not contributed to monitoring and analyzing the rhetoric of digital Arabic discourse although it has prominence, particularly in social media and has strong effectiveness in the political and social life of Arab society. This discourse has made its impact by using very new rhetorical techniques in language, voice, image, painting and video clips which are known as “Multimedia” and belong to “Digital Rhetoric”. This study suggests that it is time to draw the attention of Arab scholars and invite them to monitor and analyze the rhetoric of digital Arabic discourse.
The purpose of this study is to investigate Alahsa Governorate status and its vulnerability to generate urban heat islands. Alahsa Governorate is a famous oasis in the Arabic Peninsula including several oil centers. Extensive literature review was done to collect previous relative data on the urban heat island of Alahsa Governorate. Data used for the purpose of this research were collected from authorized bodies who control weather station networks over Alahsa Governorate, Eastern Province, Saudi Arabia. Although, the number of weather station networks within the region is very limited and the analysis using GIS software and its techniques is difficult and limited, the data analyzed confirm an increase in temperature for more than 2 °C from 2004 to 2014. Such increase is considerable whenever human health and comfort are the concern. The increase of temperature within one decade confirms the availability of urban heat islands. The study concludes that, Alahsa Governorate is vulnerable to create urban heat islands and more attention should be drawn to strategic planning of the governorate that is developing with a high pace and considerable increasing levels of urbanization.
This study will provide a brief description of the stress in Najdi Arabic dialect as well as Modern Standard Arabic. Beyond the analysis of stress patterns, this paper will also attempt to deal with two important phenomena that affect stress, namely epenthesis/insertion, vowel shortening, and consonant (the glottal stop) deletion.
Social media continues to grow, and user interfaces may become more appealing if cultural characteristics are incorporated into their design. Facebook was designed in the west, and the original language was English. Subsequently, the words in the user interface were translated to other languages, including Arabic. Arabic words are written from right to left, and English is written from left to right. The translated version may misrepresent the original design and users’ preferences may be influenced by their culture, which should be considered in the user interface design. Previous research indicates that users are more comfortable when interacting with a user interface, which relates to their own culture. Therefore, this paper, using a survey, investigates the preferences of Saudi Arabians on the Arabic version of the user interface of Facebook.
This paper reports a structured literature review of the application of Health Information Technology in developing countries, defined as the World Bank categories Low-income countries, Lower-middle-income, and Upper-middle-income countries. The aim was to identify and classify the various applications of health information technology to assess its current state in developing countries and explore potential areas of research. We offer specific analysis and application of HIT in Libya as one of the developing countries. A structured literature review was conducted using the following online databases: IEEE, Science Direct, PubMed, and Google Scholar. Publication dates were set for 2000-2013. For the PubMed search, publications in English, French, and Arabic were specified. Using a content analysis approach, 159 papers were analyzed and a total number of 26 factors were identified that affect the adoption of health information technology. Of the 2681 retrieved articles, 159 met the inclusion criteria which were carefully analyzed and classified. The implementation of health information technology across developing countries is varied. Whilst it was initially expected financial constraints would have severely limited health information technology implementation, some developing countries like India have nevertheless dominated the literature and taken the lead in conducting scientific research. Comparing the number of studies to the number of countries in each category, we found that Low-income countries and Lower-middle-income had more studies carried out than Upper-middle-income countries. However, whilst IT has been used in various sectors of the economy, the healthcare sector in developing countries is still failing to benefit fully from the potential advantages that IT can offer.
Large-scale data stream analysis has become one of the important business and research priorities lately. Social networks like Twitter and other micro-blogging platforms hold an enormous amount of data that is large in volume, velocity and variety. Extracting valuable information and trends out of these data would aid in a better understanding and decision-making. Multiple analysis techniques are deployed for English content. Moreover, one of the languages that produce a large amount of data over social networks and is least analyzed is the Arabic language. The proposed paper is a survey on the research efforts to analyze the Arabic content in Twitter focusing on the tools and methods used to extract the sentiments for the Arabic content on Twitter.
This paper presents a recognition system for isolated words like robot commands. It’s carried out by Time Delay Neural Networks; TDNN. To teleoperate a robot for specific tasks as turn, close, etc… In industrial environment and taking into account the noise coming from the machine. The choice of TDNN is based on its generalization in terms of accuracy, in more it acts as a filter that allows the passage of certain desirable frequency characteristics of speech; the goal is to determine the parameters of this filter for making an adaptable system to the variability of speech signal and to noise especially, for this the back propagation technique was used in learning phase. The approach was applied on commands pronounced in two languages separately: The French and Arabic. The results for two test bases of 300 spoken words for each one are 87%, 97.6% in neutral environment and 77.67%, 92.67% when the white Gaussian noisy was added with a SNR of 35 dB.
In the field of Quran Studies known as GHAREEB AL QURAN (The study of the meanings of strange words and structures in Holy Quran), it is difficult to distinguish some pragmatic meanings from conceptual meanings. One who wants to study this subject may need to look for a common usage between any two words or more; to understand general meaning, and sometimes may need to look for common differences between them, even if there are synonyms (word sisters).
Some of the distinguished scholars of Arabic linguistics believe that there are no synonym words, they believe in varieties of meaning and multi-context usage. Based on this viewpoint, our method was designedto look for synonyms of a word, then the differences that distinct the word and their synonyms.
There are many available books that use such a method e.g. synonyms books, dictionaries, glossaries, and some books on the interpretations of strange vocabulary of the Holy Quran, but it is difficult to look up words in these written works.
For that reason, we proposed a logical entity, which we called Differences Matrix (DM).
DM groups the synonyms words to extract the relations between them and to know the general meaning, which defines the skeleton of all word synonyms; this meaning is expressed by a word of its sisters.
In Differences Matrix, we used the sisters(words) as titles for rows and columns, and in the obtained cells we tried to define the row title (word) by using column title (her sister), so the relations between sisters appear, the expected result is well defined groups of sisters for each word. We represented the obtained results formally, and used the defined groups as a base for building the ontology of the Holy Quran synonyms.
Automatic transformation of paper documents into electronic documents requires document segmentation at the first stage. However, some parameters restrictions such as variations in character font sizes, different text line spacing, and also not uniform document layout structures altogether have made it difficult to design a general-purpose document layout analysis algorithm for many years. Thus in most previously reported methods it is inevitable to include these parameters. This problem becomes excessively acute and severe, especially in Persian/Arabic documents. Since the Persian/Arabic scripts differ considerably from the English scripts, most of the proposed methods for the English scripts do not render good results for the Persian scripts. In this paper, we present a novel parameter-free method for segmenting the Persian/Arabic document images which also works well for English scripts. This method segments the document image into maximal homogeneous regions and identifies them as texts and non-texts based on a pyramidal image structure. In other words the proposed method is capable of document segmentation without considering the character font sizes, text line spacing, and document layout structures. This algorithm is examined for 150 Arabic/Persian and English documents and document segmentation process are done successfully for 96 percent of documents.
Work is in on line Arabic character recognition and the principal motivation is to study the Arab manuscript with on line technology.
This system is a Markovian system, which one can see as like a Dynamic Bayesian Network (DBN). One of the major interests of these systems resides in the complete models training (topology and parameters) starting from training data.
Our approach is based on the dynamic Bayesian Networks formalism. The DBNs theory is a Bayesians networks generalization to the dynamic processes. Among our objective, amounts finding better parameters, which represent the links (dependences) between dynamic network variables.
In applications in pattern recognition, one will carry out the fixing of the structure, which obliges us to admit some strong assumptions (for example independence between some variables). Our application will relate to the Arabic isolated characters on line recognition using our laboratory database: NOUN. A neural tester proposed for DBN external optimization.
The DBN scores and DBN mixed are respectively 70.24% and 62.50%, which lets predict their further development; other approaches taking account time were considered and implemented until obtaining a significant recognition rate 94.79%.
The main focus of this study is investigating the interference of Arabic in the use of English prepositions by Libyan university students. Prepositions in the tests used in the study were categorized, according to their relation to Arabic, into similar Arabic and English prepositions (SAEP), dissimilar Arabic and English prepositions (DAEP), Arabic prepositions with no English counterparts (APEC), and English prepositions with no Arabic counterparts (EPAC).
The subjects of the study were the first year university students of the English department, Sabrata Faculty of Arts, Azzawia University; both males and females, and they were 100 students. The basic tool for data collection was a test of English prepositions; students are instructed to fill in the blanks with the correct prepositions and to put a zero (0) if no preposition was needed. The test was then handed to the subjects of the study.
The test was then scored and quantitative as well as qualitative results were obtained. Quantitative results indicated the number, percentages and rank order of errors in each of the categories and qualitative results indicated the nature and significance of those errors and their possible sources. Based on the obtained results the researcher could detect that students made more errors in the EPAC category than the other three categories and these errors could be attributed to the lack of knowledge of the different meanings of English prepositions. This lack of knowledge forced the students to adopt what is called the strategy of transfer.
Edward Said in his book Culture and Imperialism devotes the introduction to the Arabic translation. He claims that the fading echo of Orientalism in the Arab world is unlike the positive reflections of its counterpart elsewhere in the world. The probable reason behind his inquiry would be that the methodology Abu Deeb applied in translating Said's book contributed to the book having the limited impact which Said is referring to. The paper adds new insights to the body of theory and the effectiveness of the performance of translation from culture to culture. It presents a survey that can provide the reader with an overview of Said's Orientalism and the two Arabic translations of the book. It investigates some of the problems of translating cultural texts, more specifically translating features of Said's style.
The website developer and designer should follow usability guidelines to provide a user-friendly interface. Many guidelines and heuristics have been developed by previous studies to help both the developer and designer in this task, but E-government websites are special cases that require specialized guidelines. This paper introduces a set of 18 guidelines for evaluating the usability of e-government websites in general and Arabic e-government websites specifically, along with a check list of how to apply them. The validity and effectiveness of these guidelines were evaluated against a variety of user characteristics. The results indicated that the proposed set of guidelines can be used to identify qualitative similarities and differences with user testing and that the new set is best suited for evaluating general and e-governmental usability.
This paper aims to build an Arabic learning language tool using Flash CS4 professional software with action script 3.0 programming language, based on the Computer Aided Language Learning (CALL) material. An extra intention is to provide a primary tool and focus on learning Arabic as a second language to adults. It contains letters, words and sentences at the first stage. This includes interactive practices, which evaluates learners’ comprehension of the Arabic language. The system was examined and it was found that the language structure was correct and learners were satisfied regarding the system tools. The learners found the system tools efficient and simple to use. The paper's main conclusion illustrates that CALL can be applied without any hesitation to second language learners
Skew detection and correction (SDC) has a direct effect in efficiency and exactitude of documents’ segmentation and analysis and thus is considered as a very important step in documents’ analysis field. Skew is a major problem in documents’ analysis for every language. For Arabic/Persian document scripts this problem is more severe because of special features of these languages. In this paper an efficient and fast algorithm for Document Skew Detection (DSD) based on the concept of segmentation and Center of Gravity (COG) is proposed. This algorithm is examined for 150 Arabic/Persian and English documents and SDC process are done successfully for 93 percent of documents with error rate of less than 1°. This algorithm shows better results for English documents compared to Arabic/Persian documents. The proposed method is also represents favorable results for handwritten, printed and also complicated documents such as newspapers and journals even with very low quality and resolution.