• Title/Summary/Keyword: Intelligence Music

Search Result 96, Processing Time 0.023 seconds

Music Composition with Collaboratory AI Composers

  • Kim, Haekwang;You, Younghwan
    • Proceedings of the Korean Society of Broadcast Engineers Conference
    • /
    • 2021.06a
    • /
    • pp.23-25
    • /
    • 2021
  • This paper describes an approach of composing music with multiple AI composers. This approach enriches more the creativity space of artificial intelligence music composition than using only one composer. This paper presents a simple example with 2 different deep learning composers working together for composing one music. For the experiment, the two composers adopt the same deep learning architecture of an LSTM model trained with different data. The output of a composer is a sequence of notes. Each composer alternatively appends its output to the resulting music which is input to both the composers. Experiments compare different music generated by the proposed multiple composer approach with the traditional one composer approach.

  • PDF

Humming: Image Based Automatic Music Composition Using DeepJ Architecture (허밍: DeepJ 구조를 이용한 이미지 기반 자동 작곡 기법 연구)

  • Kim, Taehun;Jung, Keechul;Lee, Insung
    • Journal of Korea Multimedia Society
    • /
    • v.25 no.5
    • /
    • pp.748-756
    • /
    • 2022
  • Thanks to the competition of AlphaGo and Sedol Lee, machine learning has received world-wide attention and huge investments. The performance improvement of computing devices greatly contributed to big data processing and the development of neural networks. Artificial intelligence not only imitates human beings in many fields, but also seems to be better than human capabilities. Although humans' creation is still considered to be better and higher, several artificial intelligences continue to challenge human creativity. The quality of some creative outcomes by AI is as good as the real ones produced by human beings. Sometimes they are not distinguishable, because the neural network has the competence to learn the common features contained in big data and copy them. In order to confirm whether artificial intelligence can express the inherent characteristics of different arts, this paper proposes a new neural network model called Humming. It is an experimental model that combines vgg16, which extracts image features, and DeepJ's architecture, which excels in creating various genres of music. A dataset produced by our experiment shows meaningful and valid results. Different results, however, are produced when the amount of data is increased. The neural network produced a similar pattern of music even though it was a different classification of images, which was not what we were aiming for. However, these new attempts may have explicit significance as a starting point for feature transfer that will be further studied.

Implementation of Lighting Technique and Music Therapy for Improving Degree of Students Concentration During Lectures

  • Han, ChangPyoung;Hong, YouSik
    • International Journal of Internet, Broadcasting and Communication
    • /
    • v.12 no.3
    • /
    • pp.116-124
    • /
    • 2020
  • The advantage of the distance learning universities based on the 4th Industrial Revolution is that anyone can conveniently take lectures anytime, anywhere on the web. In addition, research has been actively conducted on the effect of light color and temperature control upon student performance during online classes. However, research on how the conditions of subjects, lighting colors, and music selection improve the degree of a student's concentration during online lectures has not been completed. To solve these problems in this paper, we have developed automatic analysis system SW for the weak subjects of learners by applying intelligent analysis algorithm, have proposed and simulated music therapy and art therapy. Moreover, It proposed in this paper an algorithm for an automatic analysis system, which shows the weak subjects of learners by adopting intelligence analysis algorithms. We also have presented and simulated a music therapy and art therapy algorithms, based on the blended learning, in order to increase students concentration during lecture.

Music Generation Algorithm based on the Color-Emotional Effect of a Painting (그림의 색채 감정 효과를 기반으로 한 음악 생성 알고리즘)

  • Choi, Hee Ju;Hwang, Jung-Hun;Ryu, Shinhye;Kim, Sangwook
    • Journal of Korea Multimedia Society
    • /
    • v.23 no.6
    • /
    • pp.765-771
    • /
    • 2020
  • To enable AI(artificial intelligence) to realize visual emotions, it attempts to create music centered on color, an element that causes emotions in paintings. Traditional image-based music production studies have a limitation in playing notes that are unrelated to the picture because of the absence of musical elements. In this paper, we propose a new algorithm to set the group of music through the average color of the picture, and to produce music after adding diatonic code progression and deleting sound using median value. And the results obtained through the proposed algorithm were analyzed.

Comparative Analysis of and Future Directions for AI-Based Music Composition Programs (인공지능 기반 작곡 프로그램의 비교분석과 앞으로 나아가야 할 방향에 관하여)

  • Eun Ji Park
    • The Journal of the Convergence on Culture Technology
    • /
    • v.9 no.4
    • /
    • pp.309-314
    • /
    • 2023
  • This study examines the development and limitations of current artificial intelligence (AI) music composition programs. AI music composition programs have progressed significantly owing to deep learning technology. However, they possess limitations pertaining to the creative aspects of music. In this study, we collect, compare, and analyze information on existing AI-based music composition programs and explore their technical orientation, musical concept, and drawbacks to delineate future directions for AI music composition programs. Furthermore, this study emphasizes the importance of developing AI music composition programs that create "personalized" music, aligning with the era of personalization. Ultimately, for AI-based composition programs, it is critical to extensively research how music, as an output, can touch the listeners and implement appropriate changes. By doing so, AI-based music composition programs are expected to form a new structure in and advance the music industry.

Korean Traditional Music Melody Generator using Artificial Intelligence (인공지능을 이용한 국악 멜로디 생성기에 관한 연구)

  • Bae, Jun
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.25 no.7
    • /
    • pp.869-876
    • /
    • 2021
  • In the field of music, various AI composition methods using machine learning have recently been attempted. However, most of this research has been centered on Western music, and little research has been done on Korean traditional music. Therefore, in this paper, we will create a data set of Korean traditional music, create a melody using three algorithms based on the data set, and compare the results. Three models were selected based on the similarity between language and music, LSTM, Music Transformer and Self Attention. Using each of the three models, a melody generator was modeled and trained to generate melodies. As a result of user evaluation, the Self Attention method showed higher preference than the other methods. Data set is very important in AI composition. For this, a Korean traditional music data set was created, and AI composition was attempted with various algorithms, and this is expected to be helpful in future research on AI composition for Korean traditional music.

A Study on the Performance Improvement of MLP Model for Kodály Hand Sign Scale Recognition

  • Na Gyeom YANG;Dong Kun CHUNG
    • Korean Journal of Artificial Intelligence
    • /
    • v.12 no.3
    • /
    • pp.33-39
    • /
    • 2024
  • In this paper, we explore the application of Kodaly hand signs in enhancing children's music education, performances, and auditory assistance technologies. This research focuses on improving the recognition rate of Multilayer Perceptron (MLP) models in identifying Kodaly hand sign scales through the integration of Artificial Neural Networks (ANN). We developed an enhanced MLP model by augmenting it with additional parameters and optimizing the number of hidden layers, aiming to substantially increase the model's accuracy and efficiency. The augmented model demonstrated a significant improvement in recognizing complex hand sign sequences, achieving a higher accuracy compared to previous methods. These advancements suggest that our approach can greatly benefit music education and the development of auditory assistance technologies by providing more reliable and precise recognition of Kodaly hand signs. This study confirms the potential of parameter augmentation and hidden layers optimization in refining the capabilities of neural network models for practical applications.

Musical Genre Classification Based on Deep Residual Auto-Encoder and Support Vector Machine

  • Xue Han;Wenzhuo Chen;Changjian Zhou
    • Journal of Information Processing Systems
    • /
    • v.20 no.1
    • /
    • pp.13-23
    • /
    • 2024
  • Music brings pleasure and relaxation to people. Therefore, it is necessary to classify musical genres based on scenes. Identifying favorite musical genres from massive music data is a time-consuming and laborious task. Recent studies have suggested that machine learning algorithms are effective in distinguishing between various musical genres. However, meeting the actual requirements in terms of accuracy or timeliness is challenging. In this study, a hybrid machine learning model that combines a deep residual auto-encoder (DRAE) and support vector machine (SVM) for musical genre recognition was proposed. Eight manually extracted features from the Mel-frequency cepstral coefficients (MFCC) were employed in the preprocessing stage as the hybrid music data source. During the training stage, DRAE was employed to extract feature maps, which were then used as input for the SVM classifier. The experimental results indicated that this method achieved a 91.54% F1-score and 91.58% top-1 accuracy, outperforming existing approaches. This novel approach leverages deep architecture and conventional machine learning algorithms and provides a new horizon for musical genre classification tasks.

Adaptive Beamforming System Architecture Based on AOA Estimator (AOA 추정기 기반의 적응 빔형성 시스템 구조)

  • Mun, Ji-Youn;Bae, Young-Chul;Hwang, Suk-Seung
    • The Journal of the Korea institute of electronic communication sciences
    • /
    • v.12 no.5
    • /
    • pp.777-782
    • /
    • 2017
  • The Signal Intelligence (SIGINT) system based on the adaptive beamformer, comprised of the AOA estimator followed by the interference canceller, is a cutting edge technology for collecting various signal information utilizing all sorts of devices such as the radar and satellite. In this paper, we present the efficient adaptive SIGINT structure consisted of an AOA estimator and an adaptive beamformer. For estimating AOA information of various signals, we employ the Multiple Signal Classification (MUSIC) algorithm and for efficiently suppressing high-power interference signals, we employ the Minimum Variance Distortionless Response (MVDR) algorithm. Also, we provide computer simulation examples to verify the performance of the presented adaptive beamformer structure.

An Embedding /Extracting Method of Audio Watermark Information for High Quality Stereo Music (고품질 스테레오 음악을 위한 오디오 워터마크 정보 삽입/추출 기술)

  • Bae, Kyungyul
    • Journal of Intelligence and Information Systems
    • /
    • v.24 no.2
    • /
    • pp.21-35
    • /
    • 2018
  • Since the introduction of MP3 players, CD recordings have gradually been vanishing, and the music consuming environment of music users is shifting to mobile devices. The introduction of smart devices has increased the utilization of music through music playback, mass storage, and search functions that are integrated into smartphones and tablets. At the time of initial MP3 player supply, the bitrate of the compressed music contents generally was 128 Kbps. However, as increasing of the demand for high quality music, sound quality of 384 Kbps appeared. Recently, music content of FLAC (Free License Audio Codec) format using lossless compression method is becoming popular. The download service of many music sites in Korea has classified by unlimited download with technical protection and limited download without technical protection. Digital Rights Management (DRM) technology is used as a technical protection measure for unlimited download, but it can only be used with authenticated devices that have DRM installed. Even if music purchased by the user, it cannot be used by other devices. On the contrary, in the case of music that is limited in quantity but not technically protected, there is no way to enforce anyone who distributes it, and in the case of high quality music such as FLAC, the loss is greater. In this paper, the author proposes an audio watermarking technology for copyright protection of high quality stereo music. Two kinds of information, "Copyright" and "Copy_free", are generated by using the turbo code. The two watermarks are composed of 9 bytes (72 bits). If turbo code is applied for error correction, the amount of information to be inserted as 222 bits increases. The 222-bit watermark was expanded to 1024 bits to be robust against additional errors and finally used as a watermark to insert into stereo music. Turbo code is a way to recover raw data if the damaged amount is less than 15% even if part of the code is damaged due to attack of watermarked content. It can be extended to 1024 bits or it can find 222 bits from some damaged contents by increasing the probability, the watermark itself has made it more resistant to attack. The proposed algorithm uses quantization in DCT so that watermark can be detected efficiently and SNR can be improved when stereo music is converted into mono. As a result, on average SNR exceeded 40dB, resulting in sound quality improvements of over 10dB over traditional quantization methods. This is a very significant result because it means relatively 10 times improvement in sound quality. In addition, the sample length required for extracting the watermark can be extracted sufficiently if the length is shorter than 1 second, and the watermark can be completely extracted from music samples of less than one second in all of the MP3 compression having a bit rate of 128 Kbps. The conventional quantization method can extract the watermark with a length of only 1/10 compared to the case where the sampling of the 10-second length largely fails to extract the watermark. In this study, since the length of the watermark embedded into music is 72 bits, it provides sufficient capacity to embed necessary information for music. It is enough bits to identify the music distributed all over the world. 272 can identify $4*10^{21}$, so it can be used as an identifier and it can be used for copyright protection of high quality music service. The proposed algorithm can be used not only for high quality audio but also for development of watermarking algorithm in multimedia such as UHD (Ultra High Definition) TV and high-resolution image. In addition, with the development of digital devices, users are demanding high quality music in the music industry, and artificial intelligence assistant is coming along with high quality music and streaming service. The results of this study can be used to protect the rights of copyright holders in these industries.