• Title/Summary/Keyword: file

Search Result 4,410, Processing Time 0.031 seconds

Effects of ta-C Coatings on Surface Characteristics of Dental Ni-Ti Files (치과용 Ni-Ti파일의 표면특성에 미치는 ta-C코팅효과)

  • Sun-Kyun Park;Han-Cheol Choe
    • Corrosion Science and Technology
    • /
    • v.22 no.5
    • /
    • pp.368-376
    • /
    • 2023
  • Dental Ni-Ti files must ensure stability and resistance to fatigue fracture. DLC and ta-C were coated to remove defects on the surface and ensure stability, and the surface characteristics were investigated. When coated with DLC, it was black, and in case of ta-C coating, it was blue-black. Scratches, which are defects caused by mechanical processing, were formed on the surface of the un-coated Ni-Ti file from the end of the file along the direction of processing, with the Pro-file appearing in the vertical direction and the K-file appearing in the file direction. Scratches were reduced on the coated surface, and the surface roughness was greatly reduced after coating compared to before coating. The un-coated Ni-Ti file had the lowest hardness, the DLC-coated file had the highest hardness, and ta-C showed relatively high hardness. The elastic modulus of the DLC coating film was high, and the ta-C elastic modulus was low. The adhesion of the DLC coating film tended to be higher than that of ta-C, and the wear loss amount of DLC coating of taC was lower. The corrosion potential of the ta-C coating increased significantly, and the corrosion current density decreased.

Design of System for Avoiding upload of Identical-file using SA Hash Algorithm (SA 해쉬 알고리즘을 이용한 중복파일 업로드 방지 시스템 설계)

  • Hwang, Sung-Min;Kim, Seog-Gyu
    • Journal of the Korea Society of Computer and Information
    • /
    • v.19 no.10
    • /
    • pp.81-89
    • /
    • 2014
  • In this paper, we propose SA hash algorithm to avoid upload identical files and design server system using proposed SA hash algorithm. Client to want to upload file examines the value of SA hash and if the same file is found in server system client use the existing file without upload. SA hash algorithm which is able to examine the identical-file divides original file into blocks of n bits. Original file's mod i bit and output hash value's i bit is calculated with XOR operation. It is SA hash algorithm's main routine to repeat the calculation with XOR until the end of original file. Using SA hash algorithm which is more efficient than MD5, SHA-1 and SHA-2, we can design server system to avoid uploading identical file and save storage capacity and upload-time of server system.

Analyzing the Overhead of the Memory Mapped File I/O for In-Memory File Systems (메모리 파일시스템에서 메모리 매핑을 이용한 파일 입출력의 오버헤드 분석)

  • Choi, Jungsik;Han, Hwansoo
    • KIISE Transactions on Computing Practices
    • /
    • v.22 no.10
    • /
    • pp.497-503
    • /
    • 2016
  • Emerging next-generation storage technologies such as non-volatile memory will help eliminate almost all of the storage latency that has plagued previous storage devices. In conventional storage systems, the latency of slow storage devices dominates access latency; hence, software efficiency is not critical. With low-latency storage, software costs can quickly dominate memory latency. Hence, researchers have proposed the memory mapped file I/O to avoid the software overhead. Mapping a file into the user memory space enables users to access the file directly. Therefore, it is possible to avoid the complicated I/O stack. This minimizes the number of user/kernel mode switchings. In addition, there is no data copy between kernel and user areas. Despite of the benefits in the memory mapped file I/O, its overhead still needs to be addressed, as the existing mechanism for the memory mapped file I/O is designed for slow block devices. In this paper, we identify the overheads of the memory mapped file I/O via experiments.

QuTFTP: Quick UDP Trivial File Transfer Protocol (QuTFTP: UDP 기반의 빠른 파일전송)

  • Kim, Byoung-Kug
    • Journal of Advanced Navigation Technology
    • /
    • v.24 no.5
    • /
    • pp.438-443
    • /
    • 2020
  • To transfer files between nodes on network based on Ethernet, file transfer protocol (FTP) on TCP/IP and trivial file transfer protocol (TFTP) on UDP/IP are mostly used. Due to the lack of resources (processor, memory and so on) in the embedded system where we generally use for simple works with small firmware like ones; many of the systems implement only UDP/IP for their network stacks. Thus, TFTP is greatly to be preferred. For examples, environmental sensor devices for sensor networks, Boot Loader for general embedded device and preboot execution environment (PXE) boot for PC provide the TFTP. The logic of TFTP is simple for file transmission but, there is Stop-And-Wait problem during the process which occurs long blocking time. In this paper, we propose an algorithm which called QuTFTP(Quick UDP Trivial File Transfer Protocol) to reduce the length of the blocking time and to be compatible with the legacy TFTP.

An Efficient Cleaning Scheme for File Defragmentation on Log-Structured File System (로그 구조 파일 시스템의 파일 단편화 해소를 위한 클리닝 기법)

  • Park, Jonggyu;Kang, Dong Hyun;Seo, Euiseong;Eom, Young Ik
    • Journal of KIISE
    • /
    • v.43 no.6
    • /
    • pp.627-635
    • /
    • 2016
  • When many processes issue write operations alternately on Log-structured File System (LFS), the created files can be fragmented on the file system layer although LFS sequentially allocates new blocks of each process. Unfortunately, this file fragmentation degrades read performance because it increases the number of block I/Os. Additionally, read-ahead operations which increase the number of data to request at a time exacerbates the performance degradation. In this paper, we suggest a new cleaning method on LFS that minimizes file fragmentation. During a cleaning process of LFS, our method sorts valid data blocks by inode numbers before copying the valid blocks to a new segment. This sorting re-locates fragmented blocks contiguously. Our cleaning method experimentally eliminates 60% of file fragmentation as compared to file fragmentation before cleaning. Consequently, our cleaning method improves sequential read throughput by 21% when read-ahead is applied.

A Study on the Standardization of JPEG Format as a Long-Term Preservation Master File for Paper Archives in the Central Archives of Korea (중앙기록물관리기관의 종이기록물 영구보존용 마스터 파일로서 JPEG 포맷의 표준화에 대한 연구)

  • Kang, HyenMin
    • Journal of Korean Library and Information Science Society
    • /
    • v.47 no.4
    • /
    • pp.489-510
    • /
    • 2016
  • It is defined in the internal standard of the NAK namely a NAK/G8-1:2013(v.1.) such as TIFF, PNG, PDF/A, JPEG 2000 as a Long-Term Preservation Master File for internal Standardization of Paper Archives in the Central Archives of Korea. This study reviewed the possibility that JPEG format could be a standard as a Long-Term Preservation Master File for internal Standardization of Paper Archives in the Central Archives of Korea. I pointed to the unique characteristics of JPEG file format and analysed and compared with TIFF format that is a representative master file. Through this study, I suggested that JPEG format also should be additionally included in the standard, as a master file, because it meets a qualification and functional requirements of master file and can guarantee the authenticity, reliability, integrity and usability. The meaning of this study is that JPEG format, be added to master file standard, can set up a environment that has lots of benefit in the domain of usability, compatibility to user as well as the authenticity, reliability, integrity.

A Method of Selecting Layered File System Based on Learning Block I/O History for Service-Customized Container (서비스 맞춤형 컨테이너를 위한 블록 입출력 히스토리 학습 기반 컨테이너 레이어 파일 시스템 선정 기법)

  • Yong, Chanho;Na, Sang-Ho;Lee, Pill-Woo;Huh, Eui-Nam
    • KIPS Transactions on Computer and Communication Systems
    • /
    • v.6 no.10
    • /
    • pp.415-420
    • /
    • 2017
  • Virtualization technique of OS-level is a new paradigm for deploying applications, and is attracting attention as a technology to replace traditional virtualization technique, VM (Virtual Machine). Especially, docker containers are capable of distributing application images faster and more efficient than before by applying layered image structures and union mount point to existing linux container. These characteristics of containers can only be used in layered file systems that support snapshot functionality, so it is required to select appropriate layered file systems according to the characteristics of the containerized application. We examine the characteristics of representative layered file systems and conduct write performance evaluations of each layered file systems according to the operating principles of the layered file system, Allocate-on-Demand and Copy-up. We also suggest the method of determining a appropriate layered file system principle for unknown containerized application by learning block I/O usage history of each layered file system principles in artificial neural network. Finally we validate effectiveness of artificial neural network created from block I/O history of each layered file system principles.

A Secure Deletion Method for NAND Flash File System (NAND 플래시 파일 시스템을 위한 안전 삭제 기법)

  • Lee, Jae-Heung;Oh, Jin-Ha;Kim, Seok-Hyun;Yi, Sang-Ho;Heo, Jun-Young;Cho, Yoo-Kun;Hong, Ji-Man
    • Journal of KIISE:Computing Practices and Letters
    • /
    • v.14 no.3
    • /
    • pp.251-255
    • /
    • 2008
  • In most file systems, if a file is deleted, only the metadata of the file is deleted or modified and the file's data is still stored on the physical media. Some users require that deleted files no longer be accessible. This requirement is more important in embedded systems that employ flash memory as a storage medium. In this paper, we propose a secure deletion method for NAND flash file system and apply the method to YAFFS. Our method uses encryption to delete files and forces all keys of a specific file to be stored in the same block. Therefore, only one erase operation is required to securely delete a file. Our simulation results show that the amortized number of block erases is smaller than the simple encryption method. Even though we apply our method only to the YAFFS, our method can be easily applied to other NAND flash file systems.

ELiSyR: Efficient, Lightweight and Sybil-Resilient File Search in P2P Networks

  • Kim, Hyeong-S.;Jung, Eun-Jin;Yeom, Heon-Y.
    • KSII Transactions on Internet and Information Systems (TIIS)
    • /
    • v.4 no.6
    • /
    • pp.1311-1326
    • /
    • 2010
  • Peer-to-peer (P2P) networks consume the most bandwidth in the current Internet and file sharing accounts for the majority of the P2P traffic. Thus it is important for a P2P file sharing application to be efficient in bandwidth consumption. Bandwidth consumption as much as downloaded file sizes is inevitable, but those in file search and bad downloads, e.g. wrong, corrupted, or malicious file downloads, are overheads. In this paper, we target to reduce these overheads even in the presence of high volume of malicious users and their bad files. Sybil attacks are the example of such hostile environment. Sybil attacker creates a large number of identities (Sybil nodes) and unfairly influences the system. When a large portion of the system is subverted, either in terms of the number of users or the number of files shared in the system, the overheads due to the bad downloads rapidly increase. We propose ELiSyR, a file search protocol that can tolerate such a hostile environment. ELiSyR uses social networks for P2P file search and finds benign files in 71% of searches even when more than half of the users are malicious. Furthermore, ELiSyR provides similar success with less bandwidth than other general efforts against Sybil attacks. We compare our algorithm to SybilGuard, SybilLimit and EigenTrust in terms of bandwidth consumption and the likelihood of bad downloads. Our algorithm shows lower bandwidth consumption, similar chances of bad downloads and fairer distribution of computation loads than these general efforts. In return, our algorithm takes more rounds of search than them. However the time required for search is usually much less than the time required for downloads, so the delay in search is justifiable compared to the cost of bad downloads and subsequent re-search and downloads.

Design and Implementation of a System Call Interface for Supporting File Partial Encryption (파일 부분 암호화 지원을 위한 시스템 호출의 설계 및 구현에 관한 연구)

  • Seo, Hye-In;Kim, Eun-Gi
    • Journal of the Korea Institute of Information and Communication Engineering
    • /
    • v.22 no.3
    • /
    • pp.557-567
    • /
    • 2018
  • There are currently various file encryption solutions for encrypting and storing files on disk. However, the existing file encryption solutions handle encryption and decryption all at once by file or directory. In this paper, we propose a system call supporting partial encryption function of the file. The user sets the encryption information with the system call interface at a portion where encryption of the file data is desired. And then the user writes file data, the data is encrypted and stored. Also if the user sets decryption information and reads the file data, the necessary part is decrypted by applying the set information. For the proposed system call, It consists of inspection module, management module, encryption module, decryption module, and HMAC module as per required system call. And it was implemented on the Linux environment. Also the operation of implemented system call was verified on the development board, and the performance was analyzed by measuring performance speed.