International Journal of Knowledge Content Development & Technology
/
v.14
no.4
/
pp.65-84
/
2024
The study was conducted on the awareness and application of internet of things in universities libraries in Kwara State, Nigeria. the study formulated used four research questions and used eighty five (85) samples as the population using total enumerative sampling techniques. A survey method was used in undertaking the study, in which answers were sought on the level of awareness of the internet of things in universities libraries in Kwara State, the extent of application of the internet of things in universities libraries in Kwara State, the benefit of internet of things in universities libraries in Kwara State, the challenges faced in the application of internet of things in universities libraries in Kwara State. The data collected from the study were analyzed using frequency tables and percentage. The study discovered that there the students are aware of the internet of things in universities libraries in Kwara State and the benefit of internet of a things include: Device in the IoT platforms are heterogeneous and are based on different hardware platforms and networks, It gives the high level of interoperability and interconnectivity, IoT platform has sensors which detect or measure any changes in the environment to generate data that can report on their status or even interact with the environment, IoT comes with the combination of algorithms and computation, software & hardware that makes it smart and Anything can be interconnected with the global information and communication infrastructure and the study identified data interpretation problem, Lack of skilled and specialized workers, Cost and Challenges in online security as well as Software complexity are major challenges faced in the application of internet of things in universities libraries in Kwara State. In conclusion the study made some recommendations which include that: Future libraries should be equipped with new technologies and networking devices as soon as possible. As this will be essential for users and librarians to have sufficient knowledge about IOT technologies.
Spatial neighborhoods are spaces which are relate to target space. A 3D spatial query which is a function for searching spatial neighborhoods is a significant function in spatial analysis. Various methodologies have been proposed in related these studies, this study suggests an adjacent based methodology. The methodology of this paper implements topological data for represent a adjacency via using network based topological data model, then apply modifiable Dijkstra's algorithm to each topological data. Results of ordering analysis about an adjacent space from a target space were visualized and considered ways to take advantage of. Object of this paper is to implement a 3D spatial query for searching a target space with a adjacent relationship in 3D space. And purposes of this study are to 1)generate adjacency based 3D network data via network based topological data model and to 2)implement a 3D spatial query for searching spatial neighborhoods by applying Dijkstra's algorithms to these data.
The High Level Architecture(HLA) specifies a framework for interoperation between heterogeneous simulators, and Run-Time Infrastructure(RTI) is a implementation of the HLA Interface Specification. The Data Distribution Management(DDM) services, one category of IEEE 1516 HLA/RTI management services, control filters for data transmission and reception of data volume among simulators. In this paper, we propose design concept of DDM and show its implementation for light-weighted RTI. The design concept of DDM is to minimize total amount of message that each federate and a federation process generate using the rate of RTI service execution. The design of our proposed DDM follows that a data transfer mechanism is differently applied as the rate of RTI service execution. A federate usually publishes or subscribes data when it starts. The federate constantly updates the data and modifies associated regions while it continues to advance its simulation time. Therefore, the proposed DDM design provides fast update or region modification in exchange of complex publish and subscribe services. We describe how to process the proposed DDM in IEEE 1516 HLA/RTI and experiment variable scenarios while modifying region, changing overlap ratio, and increasing data volume.
LOD(Linked Open Data) is composed of RDF triples which are based on ontologies. They are identified, linked, and accessed under the principles of linked data. Publications of LOD data sets lead to the extension of LOD cloud and ultimately progress to the web of data. However, if ontologically the same things in different LOD data sets are identified by different URIs, it is difficult to figure out their sameness and to provide trustworthy links among them. To solve this problem, we suggest a Trustworthiness Improving Link Evaluation, TILE for short, technique. TILE evaluates links in 4 steps. Step 1 is to consider the inference property of syntactic elements in LOD data set and then generate RDF triples which have existed implicitly. In Step 2, TILE appoints predicates, compares their objects in triples, and then evaluates links between the subjects in the triples. In Step 3, TILE evaluates the predicates' syntactic property at the standpoints of subject description and vocabulary definition and compensates the evaluation results of Step 2. The syntactic elements considered by TILE contain RDFS, OWL, OWL2 which are recommended by W3C. Finally, TILE makes the publisher of LOD data set review the evaluation results and then decide whether to re-evaluate or finalize the links. This leads the publishers' responsibility to be reflected in the trustworthiness of links among the data published.
Korean Journal of Construction Engineering and Management
/
v.21
no.5
/
pp.55-63
/
2020
Earthwork automation has emerged as a promising technology in the construction industry, and the application of earthwork automation technology is starting from the acquisition and processing of point cloud data of the site. Point cloud data has more than a million data due to vast extent of the construction site, and the processing time of the original point cloud data is critical because it takes tens or hundreds of hours to generate a Digital Terrain Model (DTM), and enhancement of the processing time can largely impact on the efficiency of the modeling. Currently, a benchmark program (BP) is actively used for the purpose of both point cloud data processing and BIM design as an integrated program in Korea, however, there are some aspects to be modified and refined. This study modified the BP, and developed an updated program by adopting a compile-based development environment, newly designed UI/UX, and OpenGL while maintaining existing PCD processing functions, and expended compatibility of the PCD file formats. We conducted a comparative test in terms of loading speed with different number of point cloud data, and the results showed that 92 to 99% performance increase was found in the developed program. This program can be used as a foundation for the development of a program that reduces the gap between design and construction by integrating PCD and earthwork BIM functions in the future.
The Geostationary Ocean Color Imager (GOCI) data-processing system (GDPS), which is a software system for satellite data processing and analysis of the first geostationary ocean color observation satellite, has been developed concurrently with the development of th satellite. The GDPS has functions to generate level 2 and 3 oceanographic analytical data, from level 1B data that comprise the total radiance information, by programming a specialized atmospheric algorithm and oceanic analytical algorithms to the software module. The GDPS will be a multiversion system not only as a standard Korea Ocean Satellite Center(KOSC) operational system, but also as a basic GOCI data-processing system for researchers and other users. Additionally, the GDPS will be used to make the GOCI images available for distribution by satellite network, to calculate the lookup table for radiometric calibration coefficients, to divide/mosaic several region images, to analyze time-series satellite data. the developed GDPS system has satisfied the user requirement to complete data production within 30 minutes. This system is expected to be able to be an excellent tool for monitoring both long-term and short-term changes of ocean environmental characteristics.
Recommender system has become one of the most important technologies in e-commerce in these days. The ultimate reason to shop online, for many consumers, is to reduce the efforts for information search and purchase. Recommender system is a key technology to serve these needs. Many of the past studies about recommender systems have been devoted to developing and improving recommendation algorithms and collaborative filtering (CF) is known to be the most successful one. Despite its success, however, CF has several shortcomings such as cold-start, sparsity, gray sheep problems. In order to be able to generate recommendations, ordinary CF algorithms require evaluations or preference information directly from users. For new users who do not have any evaluations or preference information, therefore, CF cannot come up with recommendations (Cold-star problem). As the numbers of products and customers increase, the scale of the data increases exponentially and most of the data cells are empty. This sparse dataset makes computation for recommendation extremely hard (Sparsity problem). Since CF is based on the assumption that there are groups of users sharing common preferences or tastes, CF becomes inaccurate if there are many users with rare and unique tastes (Gray sheep problem). This study proposes a new algorithm that utilizes Social Network Analysis (SNA) techniques to resolve the gray sheep problem. We utilize 'degree centrality' in SNA to identify users with unique preferences (gray sheep). Degree centrality in SNA refers to the number of direct links to and from a node. In a network of users who are connected through common preferences or tastes, those with unique tastes have fewer links to other users (nodes) and they are isolated from other users. Therefore, gray sheep can be identified by calculating degree centrality of each node. We divide the dataset into two, gray sheep and others, based on the degree centrality of the users. Then, different similarity measures and recommendation methods are applied to these two datasets. More detail algorithm is as follows: Step 1: Convert the initial data which is a two-mode network (user to item) into an one-mode network (user to user). Step 2: Calculate degree centrality of each node and separate those nodes having degree centrality values lower than the pre-set threshold. The threshold value is determined by simulations such that the accuracy of CF for the remaining dataset is maximized. Step 3: Ordinary CF algorithm is applied to the remaining dataset. Step 4: Since the separated dataset consist of users with unique tastes, an ordinary CF algorithm cannot generate recommendations for them. A 'popular item' method is used to generate recommendations for these users. The F measures of the two datasets are weighted by the numbers of nodes and summed to be used as the final performance metric. In order to test performance improvement by this new algorithm, an empirical study was conducted using a publically available dataset - the MovieLens data by GroupLens research team. We used 100,000 evaluations by 943 users on 1,682 movies. The proposed algorithm was compared with an ordinary CF algorithm utilizing 'Best-N-neighbors' and 'Cosine' similarity method. The empirical results show that F measure was improved about 11% on average when the proposed algorithm was used
. Past studies to improve CF performance typically used additional information other than users' evaluations such as demographic data. Some studies applied SNA techniques as a new similarity metric. This study is novel in that it used SNA to separate dataset. This study shows that performance of CF can be improved, without any additional information, when SNA techniques are used as proposed. This study has several theoretical and practical implications. This study empirically shows that the characteristics of dataset can affect the performance of CF recommender systems. This helps researchers understand factors affecting performance of CF. This study also opens a door for future studies in the area of applying SNA to CF to analyze characteristics of dataset. In practice, this study provides guidelines to improve performance of CF recommender systems with a simple modification.
Proceedings of the Korean Institute of Information and Commucation Sciences Conference
/
2012.05a
/
pp.343-346
/
2012
Today, According to it widely used that third generation mobile networks service, WCDMA module development technology and its utilization is expanding increasingly and thus to IT Convergence industries is a trend that a lot of appears. This paper, OBD-II communications to bring a vehicle information use and data to an external server transfer and it was developed that vehicle status monitoring from other external devices can be system. From the various sensors inside a vehicle using the OBD-II connector read the information, after converting the user easier to see, WCDMA module using the car status monitoring system to transfer to an external data server was implemented. Developed to test the performance of the system using virtual's state vehicle simulator, which occurs in the real car that will generate the data. It was sent to the OBD-II connector to be occurred data, vehicle status monitoring system was confirmed that Data was received without error. In addition, When this data using WCDMA transmit data to an external server the same data was confirmed that it was received without error. In the future, this technology using the OBD-II vehicle IT convergence technology can be used in a wide range.
The gas hydrate exploration using seismic reflection data, the detection of BSR(Bottom Simulating Reflector) on the seismic section is the most important work flow because the BSR have been interpreted as being formed at the base of a gas hydrate zone. Usually, BSR has some dominant qualitative characteristics on seismic section i.e. Wavelet phase reversal compare to sea bottom signal, Parallel layer with sea bottom, Strong amplitude, Masking phenomenon above the BSR, Cross bedding with other geological layer. Even though a BSR can be selected on seismic section with these guidance, it is not enough to conform as being true BSR. Some other available methods for verifying the BSR with reliable analysis quantitatively i.e. Interval velocity analysis, AVO(Amplitude Variation with Offset)analysis etc. Usually, AVO analysis can be divided by three main parts. The first part is AVO analysis, the second is AVO modeling and the last is AVO inversion. AVO analysis is unique method for detecting the free gas zone on seismic section directly. Therefore it can be a kind of useful analysis method for discriminating true BSR, which might arise from an Possion ratio contrast between high velocity layer, partially hydrated sediment and low velocity layer, water saturated gas sediment. During the AVO interpretation, as the AVO response can be changed depend upon the water saturation ratio, it is confused to discriminate the AVO response of gas layer from dry layer. In that case, the AVO modeling is necessary to generate synthetic seismogram comparing with real data. It can be available to make conclusions from correspondence or lack of correspondence between the two seismograms. AVO inversion process is the method for driving a geological model by iterative operation that the result ing synthetic seismogram matches to real data seismogram wi thin some tolerance level. AVO inversion is a topic of current research and for now there is no general consensus on how the process should be done or even whether is valid for standard seismic data. Unfortunately, there are no well log data acquired from gas hydrate exploration area in Korea. Instead of that data, well log data and seismic data acquired from gas sand area located nearby the gas hydrate exploration area is used to AVO analysis, As the results of AVO modeling, type III AVO anomaly confirmed on the gas sand layer. The Castagna's equation constant value for estimating the S-wave velocity are evaluated as A=0.86190, B=-3845.14431 respectively and water saturation ratio is $50\%$. To calculate the reflection coefficient of synthetic seismogram, the Zoeppritz equation is used. For AVO inversion process, the dataset provided by Hampson-Rushell CO. is used.
KOMPSAT-3A images have been used in various kinds of applications, since its launch in 2015. However, there were limits to scientific analysis and application extensions of these data, such as vegetation index estimation, because no tool was developed to obtain the surface reflectance required for analysis of the actual land environment. The surface reflectance is a product of performing an absolute atmospheric correction or calibration. The objective of this study is to quantitatively verify the accuracy of top-of-atmosphere reflectance and surface reflectance of KOMPSAT-3A images produced from the OTB open-source extension program, performing the cross-validation with those provided by a site measurement data of RadCalNet, an international Calibration/Validation (Cal/Val) portal. Besides, surface reflectance was obtained from Landsat-8 OLI images in the same site and applied together to the cross-validation process. According to the experiment, it is proven that the top-of-atmosphere reflectance of KOMPSAT-3A images differs by up to ± 0.02 in the range of 0.00 to 1.00 compared to the mean value of the RadCalNet data corresponding to the same spectral band. Surface reflectance in KOMPSAT-3A images also showed a high degree of consistency with RadCalNet data representing the difference of 0.02 to 0.04. These results are expected to be applicable to generate the value-added products of KOMPSAT-3A images as analysisready data (ARD). The tools applied in thisstudy and the research scheme can be extended as the new implementation of each sensor model to new types of multispectral images of compact advanced satellites (CAS) for land, agriculture, and forestry and the verification method, respectively.
본 웹사이트에 게시된 이메일 주소가 전자우편 수집 프로그램이나
그 밖의 기술적 장치를 이용하여 무단으로 수집되는 것을 거부하며,
이를 위반시 정보통신망법에 의해 형사 처벌됨을 유념하시기 바랍니다.
[게시일 2004년 10월 1일]
이용약관
제 1 장 총칙
제 1 조 (목적)
이 이용약관은 KoreaScience 홈페이지(이하 “당 사이트”)에서 제공하는 인터넷 서비스(이하 '서비스')의 가입조건 및 이용에 관한 제반 사항과 기타 필요한 사항을 구체적으로 규정함을 목적으로 합니다.
제 2 조 (용어의 정의)
① "이용자"라 함은 당 사이트에 접속하여 이 약관에 따라 당 사이트가 제공하는 서비스를 받는 회원 및 비회원을
말합니다.
② "회원"이라 함은 서비스를 이용하기 위하여 당 사이트에 개인정보를 제공하여 아이디(ID)와 비밀번호를 부여
받은 자를 말합니다.
③ "회원 아이디(ID)"라 함은 회원의 식별 및 서비스 이용을 위하여 자신이 선정한 문자 및 숫자의 조합을
말합니다.
④ "비밀번호(패스워드)"라 함은 회원이 자신의 비밀보호를 위하여 선정한 문자 및 숫자의 조합을 말합니다.
제 3 조 (이용약관의 효력 및 변경)
① 이 약관은 당 사이트에 게시하거나 기타의 방법으로 회원에게 공지함으로써 효력이 발생합니다.
② 당 사이트는 이 약관을 개정할 경우에 적용일자 및 개정사유를 명시하여 현행 약관과 함께 당 사이트의
초기화면에 그 적용일자 7일 이전부터 적용일자 전일까지 공지합니다. 다만, 회원에게 불리하게 약관내용을
변경하는 경우에는 최소한 30일 이상의 사전 유예기간을 두고 공지합니다. 이 경우 당 사이트는 개정 전
내용과 개정 후 내용을 명확하게 비교하여 이용자가 알기 쉽도록 표시합니다.
제 4 조(약관 외 준칙)
① 이 약관은 당 사이트가 제공하는 서비스에 관한 이용안내와 함께 적용됩니다.
② 이 약관에 명시되지 아니한 사항은 관계법령의 규정이 적용됩니다.
제 2 장 이용계약의 체결
제 5 조 (이용계약의 성립 등)
① 이용계약은 이용고객이 당 사이트가 정한 약관에 「동의합니다」를 선택하고, 당 사이트가 정한
온라인신청양식을 작성하여 서비스 이용을 신청한 후, 당 사이트가 이를 승낙함으로써 성립합니다.
② 제1항의 승낙은 당 사이트가 제공하는 과학기술정보검색, 맞춤정보, 서지정보 등 다른 서비스의 이용승낙을
포함합니다.
제 6 조 (회원가입)
서비스를 이용하고자 하는 고객은 당 사이트에서 정한 회원가입양식에 개인정보를 기재하여 가입을 하여야 합니다.
제 7 조 (개인정보의 보호 및 사용)
당 사이트는 관계법령이 정하는 바에 따라 회원 등록정보를 포함한 회원의 개인정보를 보호하기 위해 노력합니다. 회원 개인정보의 보호 및 사용에 대해서는 관련법령 및 당 사이트의 개인정보 보호정책이 적용됩니다.
제 8 조 (이용 신청의 승낙과 제한)
① 당 사이트는 제6조의 규정에 의한 이용신청고객에 대하여 서비스 이용을 승낙합니다.
② 당 사이트는 아래사항에 해당하는 경우에 대해서 승낙하지 아니 합니다.
- 이용계약 신청서의 내용을 허위로 기재한 경우
- 기타 규정한 제반사항을 위반하며 신청하는 경우
제 9 조 (회원 ID 부여 및 변경 등)
① 당 사이트는 이용고객에 대하여 약관에 정하는 바에 따라 자신이 선정한 회원 ID를 부여합니다.
② 회원 ID는 원칙적으로 변경이 불가하며 부득이한 사유로 인하여 변경 하고자 하는 경우에는 해당 ID를
해지하고 재가입해야 합니다.
③ 기타 회원 개인정보 관리 및 변경 등에 관한 사항은 서비스별 안내에 정하는 바에 의합니다.
제 3 장 계약 당사자의 의무
제 10 조 (KISTI의 의무)
① 당 사이트는 이용고객이 희망한 서비스 제공 개시일에 특별한 사정이 없는 한 서비스를 이용할 수 있도록
하여야 합니다.
② 당 사이트는 개인정보 보호를 위해 보안시스템을 구축하며 개인정보 보호정책을 공시하고 준수합니다.
③ 당 사이트는 회원으로부터 제기되는 의견이나 불만이 정당하다고 객관적으로 인정될 경우에는 적절한 절차를
거쳐 즉시 처리하여야 합니다. 다만, 즉시 처리가 곤란한 경우는 회원에게 그 사유와 처리일정을 통보하여야
합니다.
제 11 조 (회원의 의무)
① 이용자는 회원가입 신청 또는 회원정보 변경 시 실명으로 모든 사항을 사실에 근거하여 작성하여야 하며,
허위 또는 타인의 정보를 등록할 경우 일체의 권리를 주장할 수 없습니다.
② 당 사이트가 관계법령 및 개인정보 보호정책에 의거하여 그 책임을 지는 경우를 제외하고 회원에게 부여된
ID의 비밀번호 관리소홀, 부정사용에 의하여 발생하는 모든 결과에 대한 책임은 회원에게 있습니다.
③ 회원은 당 사이트 및 제 3자의 지적 재산권을 침해해서는 안 됩니다.
제 4 장 서비스의 이용
제 12 조 (서비스 이용 시간)
① 서비스 이용은 당 사이트의 업무상 또는 기술상 특별한 지장이 없는 한 연중무휴, 1일 24시간 운영을
원칙으로 합니다. 단, 당 사이트는 시스템 정기점검, 증설 및 교체를 위해 당 사이트가 정한 날이나 시간에
서비스를 일시 중단할 수 있으며, 예정되어 있는 작업으로 인한 서비스 일시중단은 당 사이트 홈페이지를
통해 사전에 공지합니다.
② 당 사이트는 서비스를 특정범위로 분할하여 각 범위별로 이용가능시간을 별도로 지정할 수 있습니다. 다만
이 경우 그 내용을 공지합니다.
제 13 조 (홈페이지 저작권)
① NDSL에서 제공하는 모든 저작물의 저작권은 원저작자에게 있으며, KISTI는 복제/배포/전송권을 확보하고
있습니다.
② NDSL에서 제공하는 콘텐츠를 상업적 및 기타 영리목적으로 복제/배포/전송할 경우 사전에 KISTI의 허락을
받아야 합니다.
③ NDSL에서 제공하는 콘텐츠를 보도, 비평, 교육, 연구 등을 위하여 정당한 범위 안에서 공정한 관행에
합치되게 인용할 수 있습니다.
④ NDSL에서 제공하는 콘텐츠를 무단 복제, 전송, 배포 기타 저작권법에 위반되는 방법으로 이용할 경우
저작권법 제136조에 따라 5년 이하의 징역 또는 5천만 원 이하의 벌금에 처해질 수 있습니다.
제 14 조 (유료서비스)
① 당 사이트 및 협력기관이 정한 유료서비스(원문복사 등)는 별도로 정해진 바에 따르며, 변경사항은 시행 전에
당 사이트 홈페이지를 통하여 회원에게 공지합니다.
② 유료서비스를 이용하려는 회원은 정해진 요금체계에 따라 요금을 납부해야 합니다.
제 5 장 계약 해지 및 이용 제한
제 15 조 (계약 해지)
회원이 이용계약을 해지하고자 하는 때에는 [가입해지] 메뉴를 이용해 직접 해지해야 합니다.
제 16 조 (서비스 이용제한)
① 당 사이트는 회원이 서비스 이용내용에 있어서 본 약관 제 11조 내용을 위반하거나, 다음 각 호에 해당하는
경우 서비스 이용을 제한할 수 있습니다.
- 2년 이상 서비스를 이용한 적이 없는 경우
- 기타 정상적인 서비스 운영에 방해가 될 경우
② 상기 이용제한 규정에 따라 서비스를 이용하는 회원에게 서비스 이용에 대하여 별도 공지 없이 서비스 이용의
일시정지, 이용계약 해지 할 수 있습니다.
제 17 조 (전자우편주소 수집 금지)
회원은 전자우편주소 추출기 등을 이용하여 전자우편주소를 수집 또는 제3자에게 제공할 수 없습니다.
제 6 장 손해배상 및 기타사항
제 18 조 (손해배상)
당 사이트는 무료로 제공되는 서비스와 관련하여 회원에게 어떠한 손해가 발생하더라도 당 사이트가 고의 또는 과실로 인한 손해발생을 제외하고는 이에 대하여 책임을 부담하지 아니합니다.
제 19 조 (관할 법원)
서비스 이용으로 발생한 분쟁에 대해 소송이 제기되는 경우 민사 소송법상의 관할 법원에 제기합니다.
[부 칙]
1. (시행일) 이 약관은 2016년 9월 5일부터 적용되며, 종전 약관은 본 약관으로 대체되며, 개정된 약관의 적용일 이전 가입자도 개정된 약관의 적용을 받습니다.