Skip to main content
We live in the century of technology, where the enormous evolution of data and science has recently favored a strong interest in processing, transmitting, and storing information. If, in the past, only a human mind could extract... more
We live in the century of technology, where the enormous evolution of data and science has recently favored a strong interest in processing, transmitting, and storing information. If, in the past, only a human mind could extract meaningful information from image data, after decades of dedicated research, scientists have managed to build complex systems that can identify different areas, tables, and texts from scanned documents, all the obtained information being easily accessed and passed by one to another. Books, newspapers, maps, letters, drawings-all types of documents can be scanned and processed in order to become available in a digital format. In the digital world, the storage space is very small compared to physical documents, so these applications will replace millions of old paper volumes with a single memory disk and will be accessible at the same time for anyone using just Internet access and without having a risk of deterioration. Other problems, such as ecological issues, accessibility and flexibility constraints can be solved by the use of document image analysis systems. This article presents the methods and techniques used to process on-paper documents and convert them to electronic ones, starting from pixel level and getting to the level of the entire document. The main purpose of Document Image Analysis Systems is to recognize texts and graphical interpretations from images, extract, format and present their contained information accordingly to the people's needs. We will also try to provide solid ground for practitioners that implement systems from this category to enhance the unsupervised processing features in order to make physical documents easily available to the masses.
We live in a highly technologized era, where a great number of physical documents have started or must be digitized in order to make them accessible anywhere, for a great number of people. Optical Character Recognition (OCR) is one of the... more
We live in a highly technologized era, where a great number of physical documents have started or must be digitized in order to make them accessible anywhere, for a great number of people. Optical Character Recognition (OCR) is one of the techniques which are widely used in order to recognize characters from specific images obtained after scanning. Different types of systems have been developed in order to perform Optical Character Recognition for various types of documents, but the task is not easy, as documents differ not in terms of content, but have also in formats, fonts, age or deterioration. After reviewing the existing systems, the paper at hand proposes one which uses two well-known OCR engines and a voting principle based on weights. There are also analyzed the results of our combined technique, as opposed to each individual approach of the two chosen engines.
Technology advances to make life easier for people. We tend to surround us with devices as small as possible and with the highest computing power. The need for data access from everywhere is an important detail. As a consequence, digital... more
Technology advances to make life easier for people. We tend to surround us with devices as small as possible and with the highest computing power. The need for data access from everywhere is an important detail. As a consequence, digital documents have been gaining ground on printed ones and for some sectors, the latter were even replaced. The need and the obligation to preserve the written cultural heritage, represented by books and valuable documents, some of them rare and even unique, forced us to imagine a system that protects the patrimony but makes it also accessible. In order to make books easily available to the public and at the lowest possible risk for the protection of the originals, we came to the idea of designing and creating an efficient digitization system of these records. The current article presents the proposed architecture of a Document Image Analysis System that will process the information with individual modules for each type of operation. The main scope for such tool is to recognize information from the documents and extract them for electronic use. The flow of operations are indicated by user, some steps can be eliminated depending on the user's desire and needs. In order to design an efficient Document Image Analysis System, we need a 3 axis approach: Education-involving students that can receive tasks for replacing modules and validating their homework, Research-performing various tests and Performance-testing the module interconnection and enabling the system to be extremely configurable. No matter what axis is considered, the main scope is the flexibility of the system-performed by individual modules as physical binaries or collection of binaries that are linked via scripts. Each module is designed to accomplish a certain major task by executing several sub-tasks whose results, in most cases, are subject to an intelligent voting process that produces the module's output data.
The main objective of the paper is to achieve the implementation of an economical and user-friendly neurofeedback tool that can be used to treat pain, addiction, aggression, anxiety, autism, depression, Schizophrenia, epilepsy, headaches,... more
The main objective of the paper is to achieve the implementation of an economical and user-friendly neurofeedback tool that can be used to treat pain, addiction, aggression, anxiety, autism, depression, Schizophrenia, epilepsy, headaches, insomnia, Tourette syndrome, Attention Deficit Hyperactivity Disorder (ADHD), also used for the treatment of brain damage caused by stroke, trauma, and other causes in developing countries using EEG signal analysis. The analysis and the detection of the EEG can be achieved with the help of a home-made neurofeedback-built system. The technical and clinical EEG spectrum can be subdivided into-Alpha, Beta, Theta and Gamma ranges. The report will focus on the frequencies of the Beta band in accordance to the analysis of the power values and the output of the EEG analysis will be required to control the video and-or audio feedback. [22]
The Cranial electrotherapy stimulation (CES) is a medical device that can be used to treat anxiety, depression, pain and insomnia in both adult and children by passing a form of electrical stimulation to the brain. Proposed method of... more
The Cranial electrotherapy stimulation (CES) is a medical device that can be used to treat anxiety, depression, pain and insomnia in both adult and children by passing a form of electrical stimulation to the brain. Proposed method of action, side-effects, safety factors, and treatment efficacy was put into consideration and it was confirmed that CES technology has equal or greater efficacy for the treatment of anxiety, depression, pain and insomnia compared to antidepressant medications with fewer side effects. The main purpose of the research is to implement a micro-current Arduino-based low cost user-friendly Cranial Electrotherapy Stimulation tool that can be used by adult and children in underprivileged or developing countries in order to give effective treatment for Pain, Depression, Insomnia and Anxiety by analyzing with the help of an in-house designed and built system. CES has been the subject of the investigation of analysts for quite a few years. CES is a strategy that helps someone to control his or her brain waves consciously. During CES treatment, the reading is carefully recorded in a real-time manner. At that point, the signals is recorded and display to the subjects by using online feedback loop method. CES reading is a totally non-intrusive system that can be connected over and again to patients and neuro-typical children and adults with moderately no hazard or constraint. CES is known as a corresponding and effective treatment of numerous brain dysfunctions but still has its own particular advantages and disadvantages in spite of the fact that it is a non-intrusive strategy, its legitimacy has been addressed regarding indisputable logical proof. The CES have been found to have less SMR and best amplitudes than neurotypical ones with a border of specific thresholds. The proposed low-cost accurate device is a promising tool for the treatment of Pain, Depression, Insomnia and Anxiety in adult and children with poor financial resources.
Technology advances to make life easier for people. We tend to surround us with devices as small as possible and with the highest computing power. The need for data access from everywhere is an important detail. As a consequence, digital... more
Technology advances to make life easier for people. We tend to surround us with devices as small as possible and with the highest computing power. The need for data access from everywhere is an important detail. As a consequence, digital documents have been gaining ground on printed ones and for some sectors, the latter were even replaced. The need and the obligation to preserve the written cultural heritage, represented by books and valuable documents, some of them rare and even unique, forced us to imagine a system that protects the patrimony but makes it also accessible. In order to make books easily available to the public and at the lowest possible risk for the protection of the originals, we came to the idea of designing and creating an efficient digitization system of these records. The current article presents the proposed architecture of a Document Image Analysis System that will process the information with individual modules for each type of operation. The main scope for such tool is to recognize information from the documents and extract them for electronic use. The flow of operations are indicated by user, some steps can be eliminated depending on the user's desire and needs. In order to design an efficient Document Image Analysis System, we need a 3 axis approach: Education-involving students that can receive tasks for replacing modules and validating their homework, Research-performing various tests and Performance-testing the module interconnection and enabling the system to be extremely configurable. No matter what axis is considered, the main scope is the flexibility of the system-performed by individual modules as physical binaries or collection of binaries that are linked via scripts. Each module is designed to accomplish a certain major task by executing several sub-tasks whose results, in most cases, are subject to an intelligent voting process that produces the module's output data.
We live in the century of technology, where the enormous evolution of data and science has recently favored a strong interest in processing, transmitting, and storing information. If, in the past, only a human mind could extract... more
We live in the century of technology, where the enormous evolution of data and science has recently favored a strong interest in processing, transmitting, and storing information. If, in the past, only a human mind could extract meaningful information from image data, after decades of dedicated research, scientists have managed to build complex systems that can identify different areas, tables, and texts from scanned documents, all the obtained information being easily accessed and passed by one to another. Books, newspapers, maps, letters, drawings-all types of documents can be scanned and processed in order to become available in a digital format. In the digital world, the storage space is very small compared to physical documents, so these applications will replace millions of old paper volumes with a single memory disk and will be accessible at the same time for anyone using just Internet access and without having a risk of deterioration. Other problems, such as ecological issues, accessibility and flexibility constraints can be solved by the use of document image analysis systems. This article presents the methods and techniques used to process on-paper documents and convert them to electronic ones, starting from pixel level and getting to the level of the entire document. The main purpose of Document Image Analysis Systems is to recognize texts and graphical interpretations from images, extract, format and present their contained information accordingly to the people's needs. We will also try to provide solid ground for practitioners that implement systems from this category to enhance the unsupervised processing features in order to make physical documents easily available to the masses.
Indoor positioning systems are becoming a required subsystem in many ambient assisted living scenarios. Another area that would greatly benefit from the enriched context of localization is the IoT (Internet of Things) device interaction.... more
Indoor positioning systems are becoming a required subsystem in many ambient assisted living scenarios. Another area that would greatly benefit from the enriched context of localization is the IoT (Internet of Things) device interaction. However, at this moment there aren't any satisfying technologies or approaches for precise indoor positioning. This paper proposes an indoor positioning method based on trilateration using Wi-Fi RSSI measurements with corrections to eliminate the noise and the bias. The proposed method is tested in a laboratory environment and results are described.
The opportunities offered by mobile technology continue to expand with the advances in wireless broadband. These days there are a lot of entrenched business solutions, upgraded for mobile devices, which likewise have direct use to... more
The opportunities offered by mobile technology continue to expand with the advances in wireless broadband. These days there are a lot of entrenched business solutions, upgraded for mobile devices, which likewise have direct use to services of the government services, work environment practices and involvement of stakeholders. Each advance has challenges. The usage of M-Government will bring several difficulties. Amid them, the most significant for M-Government is the inclusion of infrastructure advancement, protection of information, security, legitimate issues, mobile entrance rate, availability and compatibility. Mobile WiMAX technology is an innovative wireless technology intended to empower pervasive, rapid mobile Internet access to the vastest cluster of gadgets including notebook PCs, handsets, mobile phones and buyer electronics such as gaming gadgets, cameras, and more. Mobile WiMAX conveys cheap networks and it is the first of all Internet mobile IP solution empowering proficient and scalable networks for information, video and voice. This research concerns Mobile Government powered by Mobile WiMAX technology, which delivers the benefits for Mobile Government because of high capacity, reliability, security and all IP communications targeted for the specific environment of Iraq which pose specific challenges.
This paper was designed with the main purpose of analyzing the collaborative and independent teaching approaches and their results by comparing them with other learning methods used during the Software Project Management course at the... more
This paper was designed with the main purpose of analyzing the collaborative and independent teaching approaches and their results by comparing them with other learning methods used during the Software Project Management course at the "Politehnica" University of Bucharest. Based on the experience gathered with the two aforementioned approaches we tried a "mixed" anarchic-style learning method, which is also discussed in this paper.
This paper was designed with the main purpose of analyzing the competition and collaboration based learning approaches and their results by comparing them with other learning methods used during the Software Project Management course at... more
This paper was designed with the main purpose of analyzing the competition and collaboration based learning approaches and their results by comparing them with other learning methods used during the Software Project Management course at the "Politehnica" University of Bucharest. Based on the experience gathered with the two aforementioned approaches we tried a mixed competition-collaboration learning method, which is also discussed in this paper.
Starting with the 80's, technology capable of aiding damaged human senses started to emerge, the first commonly used device being the cochlear implant. However there are other types of devices, more or less sophisticated that are not well... more
Starting with the 80's, technology capable of aiding damaged human senses started to emerge, the first commonly used device being the cochlear implant. However there are other types of devices, more or less sophisticated that are not well known. This paper investigates existing technologies able to aid persons with a visual sense deficit, by converting the stimuli and sending them for processing to a different region of the brain or through a different path.
Highlights on textured surfaces are linear combinations of diffuse and specular reflection components. It is sometimes necessary to separate these lighting elements or completely remove the specular light, especially as a preprocessing... more
Highlights on textured surfaces are linear combinations of diffuse and specular reflection components. It is sometimes necessary to separate these lighting elements or completely remove the specular light, especially as a preprocessing step for computer vision. Many methods have been proposed for separating the reflection components. The method presented in this article improves on an existing algorithm by porting it on the GPU in order to optimize the speed, using new features found in DirectX11. New test results are also offered.
The purpose of the current paper is to present an approach that can possibly improve OCR results for highly degraded documents, and also minimize dictionary based OCR post-processing. The proposed method uses k-nearest neighbors'... more
The purpose of the current paper is to present an approach that can possibly improve OCR results for highly degraded documents, and also minimize dictionary based OCR post-processing. The proposed method uses k-nearest neighbors' classification to split the input image in homogeneous regions, based on font characteristics, and then applies normalized cross correlation between word regions to find possible duplicates of the same word. These word duplicates will be collected in word lists from which a single word in each list needs to be recognizable by the OCR. All the found fragments in a word list will be replaced by the fragment of the understood word and placed back in the image.
This paper presents methods that can be used in generating an entire planet from mathematical objects, possibly starting from a small random seed. The planet will be generated only from specified mathematical objects, fractals, procedural... more
This paper presents methods that can be used in generating an entire planet from mathematical objects, possibly starting from a small random seed. The planet will be generated only from specified mathematical objects, fractals, procedural models and constructive solid geometry. This planet will include elements such as vegetation, mountains, water, waves, rocky and sandy soil and clouds.
This paper addresses the problem of Skew Detection and Correction. The problem is of critical importance in the automated content conversion systems domain, making libraries digitalization projects possible. The paper at hand consists in... more
This paper addresses the problem of Skew Detection and Correction. The problem is of critical importance in the automated content conversion systems domain, making libraries digitalization projects possible. The paper at hand consists in a comparison between the main types of skew detection algorithms and presents the reader with a study on their advantages and disadvantages as well as proposed improvements.
In this paper we would like to present and discuss a system that automatically classifies coins. This flexible system can identify coins having different features and being photographed in different light conditions. For this purpose a... more
In this paper we would like to present and discuss a system that automatically classifies coins. This flexible system can identify coins having different features and being photographed in different light conditions. For this purpose a set of strong techniques for thresholding, edge detection and frequency domain processing were used in order to generate a fingerprint as meaningful and as invariant as possible for every coin class. Usually, the capturing of digital images cannot be performed in best conditions and inconsistencies can arise due to various lighting conditions as well as the performance of the capturing device. This article is here to suggest a method to reduce problems generated by lighting, so that image characteristics are more accurate. The proposed solution improves an existing automatic coin classification algorithm by applying illumination correction before the actual classification.
One of the directions for paper document conservation is conversion to microfilms and scanned images. Since recently microfilming has been abandoned over digitization, there is a need for standards and guidelines for the conversion... more
One of the directions for paper document conservation is conversion to microfilms and scanned images. Since recently microfilming has been abandoned over digitization, there is a need for standards and guidelines for the conversion workflow. The article proposes a set of methodologies for calibrating scanning systems to ensure high quality reproduction of both microfilms and original paper prints, in terms of tonal reproduction, geometric distortion and image sharpness.
This paper presents a new conversion algorithm from color images to grayscale that attempts to overcome the drawbacks of computing the grayscale luminance value as a weighted sum of the linear-intensity values. The algorithm aims to... more
This paper presents a new conversion algorithm from color images to grayscale that attempts to overcome the drawbacks of computing the grayscale luminance value as a weighted sum of the linear-intensity values. The algorithm aims to optimize the difference between neighboring color pixels based on the "potential" luminance difference. The algorithm iteratively adjusts the values associated to each of the pixel values, so that eventually there is a relevant difference between adjacent pixels so that the features become more visible.
This paper describes a method for fast and reliable coin classification using multiple coin images, taken in various conditions by different people. It also presents a possible solution to the misclassifications when dealing with... more
This paper describes a method for fast and reliable coin classification using multiple coin images, taken in various conditions by different people. It also presents a possible solution to the misclassifications when dealing with illumination inconsistencies. It relies on automatically extracted edge and texture/color features. For edge detection, the Canny's edge detector algorithm is used, in a variant with dynamic threshold values, obtained using Otsu's threshold algorithm. In the end some experimental results are presented indicating that the proposed approach fulfills the given task effectively.
The following paper proposes a new method of transmitting video streaming data based on splitting the input video in multiframe chunks (cubes). These chunks are processed through an algorithm that applies operations of downscaling,... more
The following paper proposes a new method of transmitting video streaming data based on splitting the input video in multiframe chunks (cubes). These chunks are processed through an algorithm that applies operations of downscaling, upscaling and difference, resulting a pyramid of different resolution cubes and residues (hypercube). These residues are then transmitted to the client based on their importance. The method is adaptive to the fluctuations in the transmitting channel bandwidth and requires a minimum overhead over streaming actual video frames.
This paper will first show how range trees are represented and the basic operations that can be made on them. Using this information, some computational geometry problems will be presented where range trees can be used in order to reduce... more
This paper will first show how range trees are represented and the basic operations that can be made on them. Using this information, some computational geometry problems will be presented where range trees can be used in order to reduce the time complexity.
The aim of this paper is to offer a faster and more rigorous way for calculating the Beta-Shape. The Beta-Shape is a tight-fitting concave polygonal contour, similar to the Alpha Shape, but which remains connected when data points are... more
The aim of this paper is to offer a faster and more rigorous way for calculating the Beta-Shape. The Beta-Shape is a tight-fitting concave polygonal contour, similar to the Alpha Shape, but which remains connected when data points are farther away than the desired resolution. The proposed method uses a computational geometry approach. We also propose an improvement to the original Beta Shape, aimed at generating a more natural flowing geometry. The complexity of the original algorithm are calculated and compared to the complexity of this algorithm.
In an automatic document conversion system, which builds digital documents from scanned articles, there is the need to perform various adjustments before the scanned image is fed to the OCR system. This is because the OCR system is prone... more
In an automatic document conversion system, which builds digital documents from scanned articles, there is the need to perform various adjustments before the scanned image is fed to the OCR system. This is because the OCR system is prone to error when the text is not properly identified, aligned, de-noised, etc. Such an adjustment is the detection of page skew, an unintentional rotation of the page, which probably occurred during scanning. In this work we have used the Radon transform to detect the angle at which a page was skewed.
This project studies the possibility of developing a mobile robot that scans the environment using 3D sensors. The project also focuses on developing an algorithm that analyzes the 3D data received from the 3D sensors and runs an image... more
This project studies the possibility of developing a
mobile robot that scans the environment using 3D sensors. The
project also focuses on developing an algorithm that analyzes
the 3D data received from the 3D sensors and runs an image
stitching algorithm based on the features found in the images.
The goal is obtaining a large 3D panorama from all the
pictures that the 3D sensor has sent.
In an automatic document conversion system, which builds digital documents from scanned articles, there is a need to perform various adjustments before the scanned image is fed to the layout analysis system. This is because the layout... more
In an automatic document conversion system, which builds digital documents from scanned articles, there is a need to perform various adjustments before the scanned image is fed to the layout analysis system. This is because the layout detection system is sensitive to errors when the page elements are not properly identified, represented, denoised, etc. Such an adjustment is the detection of foreground and background or simply called a document image binarization. This paper presents a new idea for treating the common problems which may occur during the binarization phase of the documents, which considers a parameter-free local binarization algorithm which dynamically computes the window size after it sets a threshold for the standard variation value of the window. This proved to offer consistent results for a wide variety of scanned documents consisting of various old newspapers and old library documents in different languages, both handwritten and textual documents.
The main goal of this paper is to present the architecture and functionality of e-Service based platform. The project is structured along several dimensions that follow the development of complementary services, integrated to support... more
The main goal of this paper is to present the architecture and functionality of e-Service based platform. The project is structured along several dimensions that follow the development of complementary services, integrated to support everyday work experience, research and learning in the University POLITEHNICA of Bucharest (UPB). The platform support is represented by Internet as a large scale distributed environment. The current evolution of the Internet can be viewed from multiple perspectives: service oriented (Internet of Services), users centered (Internet of People), real-world integration over Internet (Internet of Things), production and use of multimedia content over Internet. The main services in eUPB 1 platform are: (i) data retrieval, aggregation and search service, (ii) communication service for heterogeneous networks, (iii) mobile services to support context-aware applications, (iv) secure data delivery infrastructure for wireless sensor networks, (v) 3DUPB-the 3D MMO virtual replica of UPB, (vi) analysis and content extraction of scanned documents, and (vii) collaboration service. This is a position paper presenting the general architecture of eUPB and a description of each services device and functionality.
Skew detection and correction is an important step in automated content conversion systems, on which overall system performance is dependent. Although there are many working solutions at the present time, the search for an algorithm that... more
Skew detection and correction is an important step in automated content conversion systems, on which overall system performance is dependent. Although there are many working solutions at the present time, the search for an algorithm that can achieve good error rates in a fast running time and on different layout types is still open, so new solutions for skew detection are needed. The paper at hand presents a neighbor clustering based approach that has the classical advantages of this class of algorithms-the speed, but delivers better accuracy, comparable with that of Hough based solutions.
We present here the results of a 3 year long nearshoring experiment, conducted in a semi-controlled academic environment. The experiment involved teams of 10-15 people from 6 European universities and proved that the biggest issue in... more
We present here the results of a 3 year long nearshoring experiment, conducted in a semi-controlled academic environment. The experiment involved teams of 10-15 people from 6 European universities and proved that the biggest issue in nearshoring lies in the difference between the 2 involved teams in terms of technical background differences.
The testing of HL7 conformance has recently gained more and more popularity due to the importance of correct intercommunication between critical systems such as e-Health applications. The HL7 messages are very complex, and there are major... more
The testing of HL7 conformance has recently gained more and more popularity due to the importance of correct intercommunication between critical systems such as e-Health applications. The HL7 messages are very complex, and there are major differences between the versions of the HL7 protocol. This paper presents our testing solution, which offers a general and extensible way of addressing these issues and others. We used a standardized testing technology, TTCN-3, and its template matching mechanism to validate the messages exchanged by medical systems. The HL7 profile we tested was QED, but our method allows extension to any other profiles.
This paper aims to present an effective method for recognising the useful print space of high-quality digital images obtained by means of automatic scanning devices.The algorithm purports to become a valid technical solution for cropping... more
This paper aims to present an effective method for
recognising the useful print space of high-quality digital images
obtained by means of automatic scanning devices.The
algorithm purports to become a valid technical solution for
cropping both single-page images, and double-page ones – in
the latter case the gutter is detected and the final result is made
of two distinct images representing the left-hand-side page and
the right-hand-side page
The problem of converting scanned documents into electronic files, especially for large electronic libraries, has been intensely studied. Document image binarization is at the base of any conversion system and requires maximum quality for... more
The problem of converting scanned documents into
electronic files, especially for large electronic libraries, has
been intensely studied. Document image binarization is at the
base of any conversion system and requires maximum quality
for the output, since it affects all subsequent processing steps.
This paper proposes a local binarization algorithm which
dynamically computes the window size based on a decrease of
the standard deviation value within the window being enlarged.
This approach proved to offer more reliable results for a wide
variety of scans of old and degraded documents.
Virtual spaces are becoming more and more present as a medium of social interaction on the internet. The biggest challenge for a massively multiuser online (MMO) application is to be able to accommodate a large number of users while... more
Virtual spaces are becoming more and more present as a medium of social interaction on the internet. The biggest challenge for a massively multiuser online (MMO) application is to be able to accommodate a large number of users while maintaining a high degree of immersion for the users of the virtual space. Current 3D MMO Servers architectures try to solve this challenge by using traditional models such client-server or peer-to-peer but suffer from limitations regarding performance, fault tolerance and cost. In this paper we propose an innovative 3D MMO Server Architecture that uses GPGPU (General Purpose programming on Graphical Processing Units) in order to solve some of these limitations.
Skew detection and correction poses particular challenges in deformed images, as well as images suffering from camera lens distortions or distortions caused by non-flat scanning surfaces. The current paper introduces an algorithm that... more
Skew detection and correction poses particular challenges in deformed images, as well as images suffering from camera lens distortions or distortions caused by non-flat scanning surfaces. The current paper introduces an algorithm that yields good results on such images and proposes a confidence measurement to estimate the accuracy of the returned skew angle.
Over the past few years, computer assisted applications have become indispensable in many domains. The medical domain does not make an exception. The need for accuracy and speed has led to collaborations between programmers and doctors in... more
Over the past few years, computer assisted applications have become indispensable in many domains. The medical domain does not make an exception. The need for accuracy and speed has led to collaborations between programmers and doctors in improving the medical activities of gathering, processing and managing patient’s data, in giving diagnoses and even in performing surgeries.
We propose a technique for medical applications, meant to reconstruct the 3D shape of the human body parts (bones, muscular tissue, etc.), and to allow doctors to remodel the 3D shape (in order to simulate a prosthesis or to simulate the resulting shape of a body part after performing surgery).
This paper describes an approach towards an orientation and skew detection for texts in scanned documents. Before using OCR systems to obtain character information from images, a preprocessing stage, comprising a number of adjustments,... more
This paper describes an approach towards an orientation and skew detection for texts in scanned documents. Before using OCR systems to obtain character information from images, a preprocessing stage, comprising a number of adjustments, has to be performed in order to obtain accurate results. One important operation that has to be considered is the skew correction, or deskewing, of the image, a fault that arises from an incorrect scanning process. This paper presents an iterative method for detecting the text orientation and skew angle, method based on histogram processing.
Covering the need for standards and guidelines in the workflow of paper document conservation through scanning, the article proposes a methodology for ensuring high quality scans in terms of image sharpness and an improved algorithm for... more
Covering the need for standards and guidelines in the
workflow of paper document conservation through scanning,
the article proposes a methodology for ensuring high quality
scans in terms of image sharpness and an improved algorithm
for computing the Modulation Transfer Function (MTF) of a
scanning system using the slanted edge method.
This paper presents a new approach for black and white conversion of greyscale images. The algorithm can be easily modified to work on colour images by changing the comparison pattern of the peaks. It uses a scanning method for black... more
This paper presents a new approach for black and white conversion of greyscale images. The algorithm can be easily modified to work on colour images by changing the comparison pattern of the peaks. It uses a scanning method for black pixel peaks and an independent threshold associated to a histogram of the scanned area. The results have been good for different types of documents.
Partitioning is a central problem in VLSI design automation, addressing circuit’s manufacturability. Circuit partitioning has multiple applications in VLSI design. One of the most common is that of dividing combinational circuits (usually... more
Partitioning is a central problem in VLSI design automation, addressing circuit’s manufacturability. Circuit partitioning has multiple applications in VLSI design. One of the most common is that of dividing combinational circuits (usually large ones) that will not fit on a single package among a number of packages. Partitioning is of practical importance for k-LUT based FPGA circuit implementation. In this work is presented new multilevel and multi-resource partitioning algorithm targeting large combinational circuits in order to efficiently use existing FPGAs circuits
Today's computer 3D graphics simulations need to provide a high degree of realism in order to create a rich and credible 3D virtual environment. Ray Tracing is a rendering method that can achieve such a degree of realism compared to the... more
Today's computer 3D graphics simulations need to provide a high degree of realism in order to create a rich and credible 3D virtual environment. Ray Tracing is a rendering method that can achieve such a degree of realism compared to the rendering mechanisms used by current graphics hardware. Until recently, Ray Tracing was not suitable to be used as a real time renderer because it is a heavy computational method and the hardware was not fast enough. In this paper we are proposing a solution to run Ray Tracing as a Multi GPGPU (General-purpose computing on graphics processing units) program and see how well Ray Tracing scales when implemented on multiple GPUs.
The research problem presented in this paper concerns the recognition of the print space in scanned images obtained through the use of automatic scanning devices. The algorithm aims to become a reliable technical solution for one of the... more
The research problem presented in this paper
concerns the recognition of the print space in scanned images
obtained through the use of automatic scanning devices. The
algorithm aims to become a reliable technical solution for one
of the most critical stages in mass document digitization
projects. Large scale tests showed that the proposed method is
effective and practical.
Document layout analysis is a process that attempts to break down the structure of a scanned document and extract important layout elements such as paragraphs, headlines, images and so on. After this process, there is a need to mark the... more
Document layout analysis is a process that attempts
to break down the structure of a scanned document and extract
important layout elements such as paragraphs, headlines,
images and so on. After this process, there is a need to mark the
identified elements through an encapsulating shape that is
visually clearly defined and easy to manipulate. This paper
describes a method for achieving this. The results are very
accurate and the method has a high degree of stability.
The most important step in automatic content conversion is the preprocessing step. Having a very good scanned document is almost a safe bet that the document will have he content extracted with a good confidence level. The current... more
The most important step in automatic content conversion is the preprocessing step. Having a very good scanned document is almost a safe bet that the document will have
he content extracted with a good confidence level.
The current paper describes some preprocessing methods which can be used in large images that must be scanned by pieces because they simply don’t fit entirely the scanner area.
This paper presents a new approach towards creating a type of hierarchy for document image page using the information given by the page separators and the Delaunay triangulation. The final result is presented under the form of a tree... more
This paper presents a new approach towards creating a type of hierarchy for document image page using the information given by the page separators and the Delaunay triangulation. The final result is presented under the form of a tree containing the information of the page in structures such as collections of pixels and using the distance between them as a binding measurement.
We developed, using structures from MV-SIS, an application dedicated to K-LUT based FPGA synthesis, named Xsynth. Main component of it, levelMap the mapping program, was implemented using the minDepth algorithm. The mapping program was... more
We developed, using structures from MV-SIS, an application dedicated to K-LUT based FPGA synthesis, named Xsynth. Main component of it, levelMap the mapping program, was implemented using the minDepth algorithm. The mapping program was instrumented in order to study and evaluate different heuristics involved in establishing best approach to find optimum delay and optimal area mapping. We did run our mapping application on many circuits from the MCNC and IWLS 2005 benchmark circuits and we obtained good results. We present our main model, procedures, measurement results and brief comparison with previous published relevant similar mapping algorithms.
Line Detection is an important problem that arises in automatic content conversion systems (among other domains). Such a system is used to generate digital documents from hard copies of books, newspapers, etc. Given a black and white... more
Line Detection is an important problem that arises in automatic content conversion systems (among other domains). Such a system is used to generate digital documents from hard copies of books, newspapers, etc. Given a black and white image, usually obtained from a scanning device, our task is to identify groups of pixels which form horizontal or vertical lines. There are various techniques used to solve this problem. In this paper we analyze a common approach (using the Hough transform), introduce a new method and compare the results.
Entity clustering is a vital feature needed by any automatic content conversion system. Such a system constructs a digital document from a hard copy of a newspaper, book, etc. At application level, the system will process an image... more
Entity clustering is a vital feature needed by any automatic content conversion system. Such a system constructs a digital document from a hard copy of a newspaper, book, etc. At application level, the system will process an image (typically black and white) and identify the various content layout elements, such as paragraphs, tables, images, columns, etc. Here is where the entity clustering mechanism comes into play. Its role is to group atomic entities (characters, points, lines) into layout elements. To achieve this, the system can take on different approaches. They mostly rely on the geometrical properties of the enclosed items, like their relative position, size, boundaries or alignment. This paper describes an approach based on 3D mesh reduction algorithms.
This paper describes a model for detecting arbitrary-form separators on a document page by using the Delaunay triangulation. The physical layouts of the documents are always hard to extract, but determining the simplest separators found... more
This paper describes a model for detecting arbitrary-form separators on a document page by using the Delaunay triangulation. The physical layouts of the documents are always hard to extract, but determining the simplest separators found in most documents is the starting point for correct layout detection.
In this paper we address the problem of binarization of scanned documents, a preprocessing requirement for most algorithms aimed at document image analysis. Two new approaches are presented, both focusing on problem areas like low... more
In this paper we address the problem of binarization of scanned documents, a preprocessing requirement for most algorithms aimed at document image analysis. Two new approaches are presented, both focusing on problem areas like low contrast, noise, and backside image showing through the paper sheet. The first proposed technique is based on an initial preprocessing step followed by the actual conversion from the continuous space to the bitonal one. The algorithm starts by enhancing document characteristics by means of contrast stretching for each color channel. Then a locally adaptive binarization process is performed, using color thresholding based on the Gaussian blur effect. The second proposed method, a conversion technique aimed at noise-removal, is performed by combining the results of a series of threshold masks. Experimental results are given in order to verify the effectiveness of the proposed algorithms.
This paper describes an approach towards obtaining a normalized measure of text resemblance in scanned images, relying on the detection of standard character features, and using a sequence of procedures and algorithms on input images, for... more
This paper describes an approach towards obtaining a normalized measure of text resemblance in scanned images, relying on the detection of standard character features, and using a sequence of procedures and algorithms on input images, for automatic content conversion purposes. The approach relies solely on geometrical characteristics of the characters, without any information regarding context or the recognition of characters.
The purpose of this paper is to present a model for white space separators detection by using simple mathematical algorithms based on distances or font measure. The algorithms present new features as they are independent of the position... more
The purpose of this paper is to present a model for
white space separators detection by using simple mathematical
algorithms based on distances or font measure. The algorithms
present new features as they are independent of the position in
page, but they start from common ideas and improve them in
order to obtain a satisfactory output.
An important problem in many content conversion applications is reconstruction (as precise as possible) of a polygonal contour bounding a given set of input points. The basic scenario is the following: considering a set of planar points... more
An important problem in many content conversion
applications is reconstruction (as precise as possible) of a
polygonal contour bounding a given set of input points. The
basic scenario is the following: considering a set of planar
points placed in a certain area, there is the need to construct a
polygon that encapsulates all the input points and reproduces
the area’s shape as fairly as possible. The main usage of this
polygon is to detach/clip the inside polygon area from the
plane, or to separate multiple areas by non-overlapping
polygons.
Large combinational circuits need to be partitioned in order to manage scarce resources on k-LUT based FPGAs. Circuit partitioning has multiple applications in VLSI design. One of the most common is that of dividing combinational circuits... more
Large combinational circuits need to be partitioned in order to manage scarce resources on k-LUT based FPGAs. Circuit partitioning has multiple applications in VLSI design. One of the most common is that of dividing combinational circuits (usually large ones) that will not fit on a single package among a number of packages. Partitioning is of practical importance for k-LUT based FPGA circuit implementation. In this work is presented multilevel multi-resource partitioning algorithm for partitioning large combinational circuits using cone-clusters in order to efficiently map existing and commercially available k-LUT based FPGAs packages. Copyright © 2002 IFAC