The Java I/O package supports Java's basic input/output system for accessing external data from sources like files and networks. It defines streams as logical entities that produce or consume information, with byte streams for binary data and character streams for Unicode text. Streams are linked to physical devices and behave consistently across different types of devices. The package defines hierarchies of input and output stream classes, with abstract base classes like InputStream/OutputStream for bytes and Reader/Writer for characters.
This document provides an overview of input/output (I/O) streams in Java, including character streams, byte streams, file streams, pipe streams, filter streams, object serialization, and random access files. It also discusses the standard input, output, and error streams in Java and how to read from and write to these streams.
File Handling Presentation As well as File handling code in java Update, Delete, Search, View, Insert in file code are available in this presentation in you face some issues so contact me 03244064060 , Also in my Email: azeemaj101@gmail.com and Twitter @azeemaj101
This document summarizes XML parsing techniques including DOM, SAX, and Microsoft XML DOM objects. DOM builds a hierarchical model of the XML document as a tree structure in memory. SAX is event-based and parses the document sequentially, triggering events. Microsoft XML DOM provides classes that map to the W3C DOM standard for manipulating XML documents. The document compares DOM and SAX, describing their advantages and disadvantages. It also outlines common DOM objects and their properties and methods for traversing and manipulating the XML document tree.
Streams are used to transfer data between a program and source/destination. They transfer data independently of the source/destination. Streams are classified as input or output streams depending on the direction of data transfer, and as byte or character streams depending on how the data is carried. Common stream classes in Java include FileInputStream, FileOutputStream, FileReader, and FileWriter for reading from and writing to files. Exceptions like FileNotFoundException may occur if a file cannot be opened.
The document discusses byte stream classes in Java. There are two types of byte streams: InputStream and OutputStream. InputStream provides methods for reading bytes of data sequentially. FileInputStream and FileOutputStream are subclasses that allow reading and writing bytes from/to files. FileInputStream can be constructed using a file path or File object and overrides InputStream methods like read() to access file bytes.
Android developer Dmitry Dogar talks on how to organize data persistence in Android using the new Room library. Topic inspired by Google Developer Group meetup.
This document provides an introduction to XML DOM (Document Object Model) including:
- XML DOM defines a standard for accessing and manipulating XML documents and is a W3C standard.
- The DOM presents an XML document as a tree structure with elements, attributes, and text as nodes.
- The DOM is separated into three levels: Core DOM, XML DOM, and HTML DOM.
- DOM properties and methods allow accessing and modifying nodes, and DOM parsing converts an XML document into accessible DOM objects.
the slide about Exception handling in java and the file and io handling in java .inbuilt java packages in for java exception.for beginner in programming
The document discusses input/output streams in Java. There are two types of streams: byte streams and character streams. Byte streams handle input and output of bytes for binary files, while character streams handle input and output of characters for text files. Java also defines three standard streams for input, output, and errors that are represented by System.in, System.out, and System.err respectively. The document provides examples of different stream types and how they are used for input and output in Java programs.
This document discusses Java I/O and streams. It begins by introducing files and the File class, which provides methods for obtaining file properties and manipulating files. It then discusses reading and writing files using byte streams like FileInputStream and FileOutputStream. Character streams like PrintWriter and BufferedReader are presented for console I/O. Other stream classes covered include buffered streams, object streams for serialization, and data streams for primitive types. The key methods of various stream classes are listed.
The java.io package contains classes for input and output in Java. It includes abstract classes like InputStream, OutputStream, Reader, and Writer as well as concrete subclasses like FileInputStream, FileOutputStream, BufferedReader, and PrintWriter. The classes use decorator patterns and handle byte streams for binary data and character streams for text. Exceptions like IOException must be caught when using these classes to ensure resources are properly closed.
In this session you will learn:
Streams
Using a stream
Manipulating the input data
Basics of the LineReader constructor
The LineWriter class
Flushing the buffer
PrintWriter
About FileDialogs
Typical FileDialog window
FileDialog constructors
Useful FileDialog methods I
Useful FileDialog methods II
Serialization
Conditions for serializability
Writing objects to a file
For more information, visit this link: https://www.mindsmapped.com/courses/software-development/online-java-training-for-beginners/
The document discusses input and output in Java using the java.io package. It provides examples of reading keyboard input using BufferedReader and reading file input using FileReader. It also provides examples of writing console output using System.out and writing to files using PrintWriter. The document explains that java.io streams provide independence from the source or destination of the input/output.
Protocol Buffers are a language-neutral, platform-neutral way of serializing structured data. They were developed at Google to address issues with encoding structured data for communication between systems. Protocol Buffers define the data structure in a .proto file, which is then used to generate code for easily reading and writing the structured data in multiple languages. They provide a smaller data size and faster parsing than XML and allow the data structure to be updated while maintaining backwards and forwards compatibility.
Part 4 of tutorials at DC2008, Berlin. (International Conference on Dublin Core and Metadata Applications). See also part 1-3 by Jane Greenberg, Pete Johnston, and Mikael Nilsson on DC history, concepts, and other schemas. This part focuses on practical issues.
The Missing Link: Metadata Conversion Workflows for Everyone
This document describes workflows developed by Utah State University and the University of Nevada, Las Vegas to streamline metadata creation between special collections and digital initiatives departments. The workflows allow for converting finding aid information into Dublin Core for uploading item records to a digital repository, and batch linking digitized content to finding aids. The processes are designed to be taught easily and performed by various staff levels to automate metadata work and make it more flexible.
The document discusses different approaches to data management and persistence in applications, including:
1) Storing objects directly in files or using a database management system (DBMS) to store data in tables while hiding physical storage details.
2) Design questions around persistence such as whether to use files, a relational or object DBMS, and how to structure the logical and physical layers.
3) Common techniques for mapping objects to relational databases like normalization, handling inheritance and associations.
4) Alternatives for designing data management classes like adding persistence methods to classes or using broker classes.
The document discusses the ADO.Net Entity Framework 4.0 and the need for object-relational mapping (ORM) tools. It covers Entity Data Modeling (EDM) components like conceptual models, storage models, and mappings. It also discusses database first and model first approaches to EDM creation. Additional topics covered include LINQ to Entities, working with stored procedures, customizing entities with T4 templates, and using POCO entities.
Metadata management for data storage spaces :
INDEXATOR is a metadata management tool that addresses the problems of organising, documenting, storing and sharing data in a research unit or infrastructure, and fits perfectly into a data management plan of a collective.
The central idea is that the storage space becomes the data repository, so the metadata should go to the data and not the other way around.
Given the diversity of domains, the approach chosen is to be both as flexible and as pragmatic as possible by allowing each collective to choose its own (controlled) vocabulary corresponding to the reality of its field and activities. The main idea is to be able to "capture" the user's metadata as easily as possible using their vocabulary. It is possible to define the whole terminology using a spreadsheet.
The choice was made for the JSON format, which is very appropriate for describing metadata, readable by both humans and machines.
This tool is built around a web interface coupled with a MongoDB database. The web interface allows you to i) Describe a dataset using metadata of various types (Description), ii) Search datasets by their metadata (Accessibility).
The data science process document outlines the typical steps involved in a data science project including: 1) setting research goals, 2) retrieving data from internal or external sources, 3) preparing data through cleansing and transformation, 4) performing exploratory data analysis, 5) building models using techniques like machine learning or statistics, and 6) presenting and automating results. It also discusses challenges in working with different file formats and the importance of understanding various formats as a data scientist.
Data binding allows linking data from a database or text file to HTML elements. It has features like on-demand content retrieval, asynchronous processing, and sorting and filtering of data. The data binding architecture consists of four main components - a data source object, data consumers, a binding agent, and a table repetition agent. It allows accessing and manipulating data from any database through a web browser.
The document discusses data binding and describes:
1) Data binding associates data from a database with HTML elements to display the data. It allows sorting and filtering of data.
2) The architecture of data binding includes data source objects, data consumers, a binding agent, and a table repetition agent.
3) Sorting and filtering of data with a tabular data control allows reordering and restricting the display of data from a CSV file.
This document provides an overview of getting started with IDA and navigating disassemblies:
- Launching IDA involves choosing a file to analyze which loads the file and displays it. The history allows reopening recent files.
- The initial analysis populates various windows like Functions and disassembles the code. Data displays include graph, text, hex, and named views.
- Navigation uses double-clicks, addresses, and the stack frame. Searches find text or binary patterns.
- Common tasks involve naming locations and variables, transforming code/data, and recognizing data structures.
Python is open source and has so many libraries for data wrangling and visualization that makes life of data scientists easier. For data wrangling pandas is used as it represent tabular data and it has other function to parse data from different sources, data cleaning, handling missing values, merging data sets etc. To visualize data, low level matplotlib can be used. But it is a base package for other high level packages such as seaborn, that draw well customized plot in just one line of code. Python has dash framework that is used to make interactive web application using python code without javascript and html. These dash application can be published on any server as well as on clouds like google cloud but freely on heroku cloud.
DataFinder concepts and example: General (20100503)
DataFinder is a lightweight client-server solution for centralized data management. It was created by the German Aerospace Center (DLR) to address the problems of absent data organization structures and no centralized policy for data management. DataFinder provides graphical user interfaces and uses a logical data store concept to organize data across distributed storage locations according to a configurable data model. It can be customized through Python scripts to integrate with different environments and automate tasks like data migration.
Putting Historical Data in Context: how to use DSpace-GLAM
This document discusses using DSpace and DSpace-GLAM to manage digital cultural heritage data. It provides an overview of DSpace's data model and functionality for ingesting, describing, and sharing digital objects. It then introduces DSpace-GLAM, an extension of DSpace developed for cultural heritage institutions. DSpace-GLAM adds additional entity types, relationships, and metadata to better represent cultural concepts. It also provides tools for visualizing and analyzing datasets.
The document discusses various technologies for metasearching or cross-searching multiple databases at once, including Z39.50 for real-time searching, SRU/SRW web services, and OAI-PMH for metadata harvesting. It explains concepts like XML, web services, SOAP, and WSDL, and provides examples of how technologies like Z39.50, SRU, and OAI-PMH enable searching across different data sources.
A brief history in TimeSeries data at Environment Canada. An Enterprise view of how FME can be integrated into departmental data management activities.
DataFinder is software developed by the German Aerospace Center (DLR) to help scientists and engineers efficiently manage and organize their large and growing scientific data sets. It provides a structured way to organize data through customizable data models and metadata, and can integrate various storage resources. DataFinder was created in Python due to its ease of use and maintainability. It uses a client-server model with a WebDAV server to manage metadata and data structures, and can access different storage backends. Customizations through Python scripts allow users to automate tasks and integrate it into their workflows.
The MIDESS Project explored sharing digital content like images between university repositories. It tested standards like OAI-PMH and METS for exchanging metadata and objects. While these standards allow some interoperability, repositories implemented them differently, preventing full sharing. The project highlighted ongoing issues around information architecture, repository functionality for multimedia, and integrating repositories into broader systems.
This document discusses requirements for designing a framework to analyze text datasets. It identifies several key variations in importing datasets related to file sources, formats and schemas. It then proposes using high-level reader classes to handle different datasets. The document outlines the STAT domain model which includes concepts like RawCorpus to represent raw document collections, Processor to process data, Corpus to represent data for machine learning, Trainer for algorithms, Model to store learned parameters, Classifier to classify documents, Prediction for output classifications, Evaluator to evaluate predictions and Evaluation for results.
DataFinder: A Python Application for Scientific Data Management
DataFinder is a Python application developed by the German Aerospace Center (DLR) for efficient management of large scientific and technical data sets. It provides a structured way to organize data through customizable data models and flexible use of distributed storage resources. DataFinder uses a client-server model with a WebDAV server to store metadata and data. It allows integration of data management into scientific workflows through a Python API and scripting.
This document provides an overview of week 4 topics for a Code Club on wrangling data with Python. The key points covered are:
- Merging data from different data frames and working with "tidy data" formats.
- Taking first steps in reshaping datasets, including long to wide data transforms and pivots.
- Various strategies for cleaning datasets, such as handling empty/duplicate rows and columns, cleaning text strings, and mapping values.
- The melt and pivot functions for reshaping data between wide and long formats.
This chapter discusses data design concepts, file processing systems, database systems, and web-based data design. It explains key data design terminology and how to draw entity relationship diagrams to represent relationships between entities. The chapter also covers database models, data storage and access methods, and data control measures to ensure security and integrity.
Understanding Insider Security Threats: Types, Examples, Effects, and Mitigat...
Today’s digitally connected world presents a wide range of security challenges for enterprises. Insider security threats are particularly noteworthy because they have the potential to cause significant harm. Unlike external threats, insider risks originate from within the company, making them more subtle and challenging to identify. This blog aims to provide a comprehensive understanding of insider security threats, including their types, examples, effects, and mitigation techniques.
To help you choose the best DiskWarrior alternative, we've compiled a comparison table summarizing the features, pros, cons, and pricing of six alternatives.
We are honored to launch and host this event for our UiPath Polish Community, with the help of our partners - Proservartner!
We certainly hope we have managed to spike your interest in the subjects to be presented and the incredible networking opportunities at hand, too!
Check out our proposed agenda below 👇👇
08:30 ☕ Welcome coffee (30')
09:00 Opening note/ Intro to UiPath Community (10')
Cristina Vidu, Global Manager, Marketing Community @UiPath
Dawid Kot, Digital Transformation Lead @Proservartner
09:10 Cloud migration - Proservartner & DOVISTA case study (30')
Marcin Drozdowski, Automation CoE Manager @DOVISTA
Pawel Kamiński, RPA developer @DOVISTA
Mikolaj Zielinski, UiPath MVP, Senior Solutions Engineer @Proservartner
09:40 From bottlenecks to breakthroughs: Citizen Development in action (25')
Pawel Poplawski, Director, Improvement and Automation @McCormick & Company
Michał Cieślak, Senior Manager, Automation Programs @McCormick & Company
10:05 Next-level bots: API integration in UiPath Studio (30')
Mikolaj Zielinski, UiPath MVP, Senior Solutions Engineer @Proservartner
10:35 ☕ Coffee Break (15')
10:50 Document Understanding with my RPA Companion (45')
Ewa Gruszka, Enterprise Sales Specialist, AI & ML @UiPath
11:35 Power up your Robots: GenAI and GPT in REFramework (45')
Krzysztof Karaszewski, Global RPA Product Manager
12:20 🍕 Lunch Break (1hr)
13:20 From Concept to Quality: UiPath Test Suite for AI-powered Knowledge Bots (30')
Kamil Miśko, UiPath MVP, Senior RPA Developer @Zurich Insurance
13:50 Communications Mining - focus on AI capabilities (30')
Thomasz Wierzbicki, Business Analyst @Office Samurai
14:20 Polish MVP panel: Insights on MVP award achievements and career profiling
Paradigm Shifts in User Modeling: A Journey from Historical Foundations to Em...
Slide of the tutorial entitled "Paradigm Shifts in User Modeling: A Journey from Historical Foundations to Emerging Trends" held at UMAP'24: 32nd ACM Conference on User Modeling, Adaptation and Personalization (July 1, 2024 | Cagliari, Italy)
Scaling Connections in PostgreSQL Postgres Bangalore(PGBLR) Meetup-2 - Mydbops
This presentation, delivered at the Postgres Bangalore (PGBLR) Meetup-2 on June 29th, 2024, dives deep into connection pooling for PostgreSQL databases. Aakash M, a PostgreSQL Tech Lead at Mydbops, explores the challenges of managing numerous connections and explains how connection pooling optimizes performance and resource utilization.
Key Takeaways:
* Understand why connection pooling is essential for high-traffic applications
* Explore various connection poolers available for PostgreSQL, including pgbouncer
* Learn the configuration options and functionalities of pgbouncer
* Discover best practices for monitoring and troubleshooting connection pooling setups
* Gain insights into real-world use cases and considerations for production environments
This presentation is ideal for:
* Database administrators (DBAs)
* Developers working with PostgreSQL
* DevOps engineers
* Anyone interested in optimizing PostgreSQL performance
Contact info@mydbops.com for PostgreSQL Managed, Consulting and Remote DBA Services
This document discusses reading and writing files in .NET applications. It describes the key classes for input and output streams, including File, Directory, StreamReader, and StreamWriter. These classes allow applications to read from and write to files to store and transfer data.
Java uses streams to handle input/output operations. Streams provide a standardized way to read from and write to various sources and sinks like files, networks, and buffers. There are byte streams that handle input/output of bytes and character streams that handle characters. Common stream classes include FileInputStream, FileOutputStream, BufferedReader, and BufferedWriter which are used to read from and write to files and console. Streams can be chained together for complex I/O processing.
Object-oriented programming Undergraduate Course Presentations
java.io streams and files in Java
University of Vale do Itajaí
Univali
Incremental Tecnologia
English version
The Java I/O package supports Java's basic input/output system for accessing external data from sources like files and networks. It defines streams as logical entities that produce or consume information, with byte streams for binary data and character streams for Unicode text. Streams are linked to physical devices and behave consistently across different types of devices. The package defines hierarchies of input and output stream classes, with abstract base classes like InputStream/OutputStream for bytes and Reader/Writer for characters.
This document provides an overview of input/output (I/O) streams in Java, including character streams, byte streams, file streams, pipe streams, filter streams, object serialization, and random access files. It also discusses the standard input, output, and error streams in Java and how to read from and write to these streams.
File Handling Presentation As well as File handling code in java Update, Delete, Search, View, Insert in file code are available in this presentation in you face some issues so contact me 03244064060 , Also in my Email: azeemaj101@gmail.com and Twitter @azeemaj101
This document summarizes XML parsing techniques including DOM, SAX, and Microsoft XML DOM objects. DOM builds a hierarchical model of the XML document as a tree structure in memory. SAX is event-based and parses the document sequentially, triggering events. Microsoft XML DOM provides classes that map to the W3C DOM standard for manipulating XML documents. The document compares DOM and SAX, describing their advantages and disadvantages. It also outlines common DOM objects and their properties and methods for traversing and manipulating the XML document tree.
Streams are used to transfer data between a program and source/destination. They transfer data independently of the source/destination. Streams are classified as input or output streams depending on the direction of data transfer, and as byte or character streams depending on how the data is carried. Common stream classes in Java include FileInputStream, FileOutputStream, FileReader, and FileWriter for reading from and writing to files. Exceptions like FileNotFoundException may occur if a file cannot be opened.
The document discusses byte stream classes in Java. There are two types of byte streams: InputStream and OutputStream. InputStream provides methods for reading bytes of data sequentially. FileInputStream and FileOutputStream are subclasses that allow reading and writing bytes from/to files. FileInputStream can be constructed using a file path or File object and overrides InputStream methods like read() to access file bytes.
Data Persistence in Android with Room LibraryReinvently
Android developer Dmitry Dogar talks on how to organize data persistence in Android using the new Room library. Topic inspired by Google Developer Group meetup.
This document provides an introduction to XML DOM (Document Object Model) including:
- XML DOM defines a standard for accessing and manipulating XML documents and is a W3C standard.
- The DOM presents an XML document as a tree structure with elements, attributes, and text as nodes.
- The DOM is separated into three levels: Core DOM, XML DOM, and HTML DOM.
- DOM properties and methods allow accessing and modifying nodes, and DOM parsing converts an XML document into accessible DOM objects.
the slide about Exception handling in java and the file and io handling in java .inbuilt java packages in for java exception.for beginner in programming
The document discusses input/output streams in Java. There are two types of streams: byte streams and character streams. Byte streams handle input and output of bytes for binary files, while character streams handle input and output of characters for text files. Java also defines three standard streams for input, output, and errors that are represented by System.in, System.out, and System.err respectively. The document provides examples of different stream types and how they are used for input and output in Java programs.
This document discusses Java I/O and streams. It begins by introducing files and the File class, which provides methods for obtaining file properties and manipulating files. It then discusses reading and writing files using byte streams like FileInputStream and FileOutputStream. Character streams like PrintWriter and BufferedReader are presented for console I/O. Other stream classes covered include buffered streams, object streams for serialization, and data streams for primitive types. The key methods of various stream classes are listed.
The java.io package contains classes for input and output in Java. It includes abstract classes like InputStream, OutputStream, Reader, and Writer as well as concrete subclasses like FileInputStream, FileOutputStream, BufferedReader, and PrintWriter. The classes use decorator patterns and handle byte streams for binary data and character streams for text. Exceptions like IOException must be caught when using these classes to ensure resources are properly closed.
In this session you will learn:
Streams
Using a stream
Manipulating the input data
Basics of the LineReader constructor
The LineWriter class
Flushing the buffer
PrintWriter
About FileDialogs
Typical FileDialog window
FileDialog constructors
Useful FileDialog methods I
Useful FileDialog methods II
Serialization
Conditions for serializability
Writing objects to a file
For more information, visit this link: https://www.mindsmapped.com/courses/software-development/online-java-training-for-beginners/
The document discusses input and output in Java using the java.io package. It provides examples of reading keyboard input using BufferedReader and reading file input using FileReader. It also provides examples of writing console output using System.out and writing to files using PrintWriter. The document explains that java.io streams provide independence from the source or destination of the input/output.
Protocol Buffers are a language-neutral, platform-neutral way of serializing structured data. They were developed at Google to address issues with encoding structured data for communication between systems. Protocol Buffers define the data structure in a .proto file, which is then used to generate code for easily reading and writing the structured data in multiple languages. They provide a smaller data size and faster parsing than XML and allow the data structure to be updated while maintaining backwards and forwards compatibility.
Part 4 of tutorials at DC2008, Berlin. (International Conference on Dublin Core and Metadata Applications). See also part 1-3 by Jane Greenberg, Pete Johnston, and Mikael Nilsson on DC history, concepts, and other schemas. This part focuses on practical issues.
The Missing Link: Metadata Conversion Workflows for EveryoneAndrea Payant
This document describes workflows developed by Utah State University and the University of Nevada, Las Vegas to streamline metadata creation between special collections and digital initiatives departments. The workflows allow for converting finding aid information into Dublin Core for uploading item records to a digital repository, and batch linking digitized content to finding aids. The processes are designed to be taught easily and performed by various staff levels to automate metadata work and make it more flexible.
The document discusses different approaches to data management and persistence in applications, including:
1) Storing objects directly in files or using a database management system (DBMS) to store data in tables while hiding physical storage details.
2) Design questions around persistence such as whether to use files, a relational or object DBMS, and how to structure the logical and physical layers.
3) Common techniques for mapping objects to relational databases like normalization, handling inheritance and associations.
4) Alternatives for designing data management classes like adding persistence methods to classes or using broker classes.
The document discusses the ADO.Net Entity Framework 4.0 and the need for object-relational mapping (ORM) tools. It covers Entity Data Modeling (EDM) components like conceptual models, storage models, and mappings. It also discusses database first and model first approaches to EDM creation. Additional topics covered include LINQ to Entities, working with stored procedures, customizing entities with T4 templates, and using POCO entities.
Metadata management for data storage spaces :
INDEXATOR is a metadata management tool that addresses the problems of organising, documenting, storing and sharing data in a research unit or infrastructure, and fits perfectly into a data management plan of a collective.
The central idea is that the storage space becomes the data repository, so the metadata should go to the data and not the other way around.
Given the diversity of domains, the approach chosen is to be both as flexible and as pragmatic as possible by allowing each collective to choose its own (controlled) vocabulary corresponding to the reality of its field and activities. The main idea is to be able to "capture" the user's metadata as easily as possible using their vocabulary. It is possible to define the whole terminology using a spreadsheet.
The choice was made for the JSON format, which is very appropriate for describing metadata, readable by both humans and machines.
This tool is built around a web interface coupled with a MongoDB database. The web interface allows you to i) Describe a dataset using metadata of various types (Description), ii) Search datasets by their metadata (Accessibility).
The data science process document outlines the typical steps involved in a data science project including: 1) setting research goals, 2) retrieving data from internal or external sources, 3) preparing data through cleansing and transformation, 4) performing exploratory data analysis, 5) building models using techniques like machine learning or statistics, and 6) presenting and automating results. It also discusses challenges in working with different file formats and the importance of understanding various formats as a data scientist.
Data binding allows linking data from a database or text file to HTML elements. It has features like on-demand content retrieval, asynchronous processing, and sorting and filtering of data. The data binding architecture consists of four main components - a data source object, data consumers, a binding agent, and a table repetition agent. It allows accessing and manipulating data from any database through a web browser.
The document discusses data binding and describes:
1) Data binding associates data from a database with HTML elements to display the data. It allows sorting and filtering of data.
2) The architecture of data binding includes data source objects, data consumers, a binding agent, and a table repetition agent.
3) Sorting and filtering of data with a tabular data control allows reordering and restricting the display of data from a CSV file.
This document provides an overview of getting started with IDA and navigating disassemblies:
- Launching IDA involves choosing a file to analyze which loads the file and displays it. The history allows reopening recent files.
- The initial analysis populates various windows like Functions and disassembles the code. Data displays include graph, text, hex, and named views.
- Navigation uses double-clicks, addresses, and the stack frame. Searches find text or binary patterns.
- Common tasks involve naming locations and variables, transforming code/data, and recognizing data structures.
Data Wrangling and Visualization Using PythonMOHITKUMAR1379
Python is open source and has so many libraries for data wrangling and visualization that makes life of data scientists easier. For data wrangling pandas is used as it represent tabular data and it has other function to parse data from different sources, data cleaning, handling missing values, merging data sets etc. To visualize data, low level matplotlib can be used. But it is a base package for other high level packages such as seaborn, that draw well customized plot in just one line of code. Python has dash framework that is used to make interactive web application using python code without javascript and html. These dash application can be published on any server as well as on clouds like google cloud but freely on heroku cloud.
DataFinder concepts and example: General (20100503)Data Finder
DataFinder is a lightweight client-server solution for centralized data management. It was created by the German Aerospace Center (DLR) to address the problems of absent data organization structures and no centralized policy for data management. DataFinder provides graphical user interfaces and uses a logical data store concept to organize data across distributed storage locations according to a configurable data model. It can be customized through Python scripts to integrate with different environments and automate tasks like data migration.
Putting Historical Data in Context: how to use DSpace-GLAM4Science
This document discusses using DSpace and DSpace-GLAM to manage digital cultural heritage data. It provides an overview of DSpace's data model and functionality for ingesting, describing, and sharing digital objects. It then introduces DSpace-GLAM, an extension of DSpace developed for cultural heritage institutions. DSpace-GLAM adds additional entity types, relationships, and metadata to better represent cultural concepts. It also provides tools for visualizing and analyzing datasets.
The document discusses various technologies for metasearching or cross-searching multiple databases at once, including Z39.50 for real-time searching, SRU/SRW web services, and OAI-PMH for metadata harvesting. It explains concepts like XML, web services, SOAP, and WSDL, and provides examples of how technologies like Z39.50, SRU, and OAI-PMH enable searching across different data sources.
Environment Canada's Data Management ServiceSafe Software
A brief history in TimeSeries data at Environment Canada. An Enterprise view of how FME can be integrated into departmental data management activities.
DataFinder is software developed by the German Aerospace Center (DLR) to help scientists and engineers efficiently manage and organize their large and growing scientific data sets. It provides a structured way to organize data through customizable data models and metadata, and can integrate various storage resources. DataFinder was created in Python due to its ease of use and maintainability. It uses a client-server model with a WebDAV server to manage metadata and data structures, and can access different storage backends. Customizations through Python scripts allow users to automate tasks and integrate it into their workflows.
The MIDESS Project explored sharing digital content like images between university repositories. It tested standards like OAI-PMH and METS for exchanging metadata and objects. While these standards allow some interoperability, repositories implemented them differently, preventing full sharing. The project highlighted ongoing issues around information architecture, repository functionality for multimedia, and integrating repositories into broader systems.
This document discusses requirements for designing a framework to analyze text datasets. It identifies several key variations in importing datasets related to file sources, formats and schemas. It then proposes using high-level reader classes to handle different datasets. The document outlines the STAT domain model which includes concepts like RawCorpus to represent raw document collections, Processor to process data, Corpus to represent data for machine learning, Trainer for algorithms, Model to store learned parameters, Classifier to classify documents, Prediction for output classifications, Evaluator to evaluate predictions and Evaluation for results.
DataFinder: A Python Application for Scientific Data ManagementAndreas Schreiber
DataFinder is a Python application developed by the German Aerospace Center (DLR) for efficient management of large scientific and technical data sets. It provides a structured way to organize data through customizable data models and flexible use of distributed storage resources. DataFinder uses a client-server model with a WebDAV server to store metadata and data. It allows integration of data management into scientific workflows through a Python API and scripting.
This document provides an overview of week 4 topics for a Code Club on wrangling data with Python. The key points covered are:
- Merging data from different data frames and working with "tidy data" formats.
- Taking first steps in reshaping datasets, including long to wide data transforms and pivots.
- Various strategies for cleaning datasets, such as handling empty/duplicate rows and columns, cleaning text strings, and mapping values.
- The melt and pivot functions for reshaping data between wide and long formats.
This chapter discusses data design concepts, file processing systems, database systems, and web-based data design. It explains key data design terminology and how to draw entity relationship diagrams to represent relationships between entities. The chapter also covers database models, data storage and access methods, and data control measures to ensure security and integrity.
Understanding Insider Security Threats: Types, Examples, Effects, and Mitigat...Bert Blevins
Today’s digitally connected world presents a wide range of security challenges for enterprises. Insider security threats are particularly noteworthy because they have the potential to cause significant harm. Unlike external threats, insider risks originate from within the company, making them more subtle and challenging to identify. This blog aims to provide a comprehensive understanding of insider security threats, including their types, examples, effects, and mitigation techniques.
Comparison Table of DiskWarrior Alternatives.pdfAndrey Yasko
To help you choose the best DiskWarrior alternative, we've compiled a comparison table summarizing the features, pros, cons, and pricing of six alternatives.
UiPath Community Day Kraków: Devs4Devs ConferenceUiPathCommunity
We are honored to launch and host this event for our UiPath Polish Community, with the help of our partners - Proservartner!
We certainly hope we have managed to spike your interest in the subjects to be presented and the incredible networking opportunities at hand, too!
Check out our proposed agenda below 👇👇
08:30 ☕ Welcome coffee (30')
09:00 Opening note/ Intro to UiPath Community (10')
Cristina Vidu, Global Manager, Marketing Community @UiPath
Dawid Kot, Digital Transformation Lead @Proservartner
09:10 Cloud migration - Proservartner & DOVISTA case study (30')
Marcin Drozdowski, Automation CoE Manager @DOVISTA
Pawel Kamiński, RPA developer @DOVISTA
Mikolaj Zielinski, UiPath MVP, Senior Solutions Engineer @Proservartner
09:40 From bottlenecks to breakthroughs: Citizen Development in action (25')
Pawel Poplawski, Director, Improvement and Automation @McCormick & Company
Michał Cieślak, Senior Manager, Automation Programs @McCormick & Company
10:05 Next-level bots: API integration in UiPath Studio (30')
Mikolaj Zielinski, UiPath MVP, Senior Solutions Engineer @Proservartner
10:35 ☕ Coffee Break (15')
10:50 Document Understanding with my RPA Companion (45')
Ewa Gruszka, Enterprise Sales Specialist, AI & ML @UiPath
11:35 Power up your Robots: GenAI and GPT in REFramework (45')
Krzysztof Karaszewski, Global RPA Product Manager
12:20 🍕 Lunch Break (1hr)
13:20 From Concept to Quality: UiPath Test Suite for AI-powered Knowledge Bots (30')
Kamil Miśko, UiPath MVP, Senior RPA Developer @Zurich Insurance
13:50 Communications Mining - focus on AI capabilities (30')
Thomasz Wierzbicki, Business Analyst @Office Samurai
14:20 Polish MVP panel: Insights on MVP award achievements and career profiling
Paradigm Shifts in User Modeling: A Journey from Historical Foundations to Em...Erasmo Purificato
Slide of the tutorial entitled "Paradigm Shifts in User Modeling: A Journey from Historical Foundations to Emerging Trends" held at UMAP'24: 32nd ACM Conference on User Modeling, Adaptation and Personalization (July 1, 2024 | Cagliari, Italy)
Scaling Connections in PostgreSQL Postgres Bangalore(PGBLR) Meetup-2 - MydbopsMydbops
This presentation, delivered at the Postgres Bangalore (PGBLR) Meetup-2 on June 29th, 2024, dives deep into connection pooling for PostgreSQL databases. Aakash M, a PostgreSQL Tech Lead at Mydbops, explores the challenges of managing numerous connections and explains how connection pooling optimizes performance and resource utilization.
Key Takeaways:
* Understand why connection pooling is essential for high-traffic applications
* Explore various connection poolers available for PostgreSQL, including pgbouncer
* Learn the configuration options and functionalities of pgbouncer
* Discover best practices for monitoring and troubleshooting connection pooling setups
* Gain insights into real-world use cases and considerations for production environments
This presentation is ideal for:
* Database administrators (DBAs)
* Developers working with PostgreSQL
* DevOps engineers
* Anyone interested in optimizing PostgreSQL performance
Contact info@mydbops.com for PostgreSQL Managed, Consulting and Remote DBA Services
Measuring the Impact of Network Latency at TwitterScyllaDB
Widya Salim and Victor Ma will outline the causal impact analysis, framework, and key learnings used to quantify the impact of reducing Twitter's network latency.
7 Most Powerful Solar Storms in the History of Earth.pdfEnterprise Wired
Solar Storms (Geo Magnetic Storms) are the motion of accelerated charged particles in the solar environment with high velocities due to the coronal mass ejection (CME).
Details of description part II: Describing images in practice - Tech Forum 2024BookNet Canada
This presentation explores the practical application of image description techniques. Familiar guidelines will be demonstrated in practice, and descriptions will be developed “live”! If you have learned a lot about the theory of image description techniques but want to feel more confident putting them into practice, this is the presentation for you. There will be useful, actionable information for everyone, whether you are working with authors, colleagues, alone, or leveraging AI as a collaborator.
Link to presentation recording and transcript: https://bnctechforum.ca/sessions/details-of-description-part-ii-describing-images-in-practice/
Presented by BookNet Canada on June 25, 2024, with support from the Department of Canadian Heritage.
TrustArc Webinar - 2024 Data Privacy Trends: A Mid-Year Check-InTrustArc
Six months into 2024, and it is clear the privacy ecosystem takes no days off!! Regulators continue to implement and enforce new regulations, businesses strive to meet requirements, and technology advances like AI have privacy professionals scratching their heads about managing risk.
What can we learn about the first six months of data privacy trends and events in 2024? How should this inform your privacy program management for the rest of the year?
Join TrustArc, Goodwin, and Snyk privacy experts as they discuss the changes we’ve seen in the first half of 2024 and gain insight into the concrete, actionable steps you can take to up-level your privacy program in the second half of the year.
This webinar will review:
- Key changes to privacy regulations in 2024
- Key themes in privacy and data governance in 2024
- How to maximize your privacy program in the second half of 2024
Implementations of Fused Deposition Modeling in real worldEmerging Tech
The presentation showcases the diverse real-world applications of Fused Deposition Modeling (FDM) across multiple industries:
1. **Manufacturing**: FDM is utilized in manufacturing for rapid prototyping, creating custom tools and fixtures, and producing functional end-use parts. Companies leverage its cost-effectiveness and flexibility to streamline production processes.
2. **Medical**: In the medical field, FDM is used to create patient-specific anatomical models, surgical guides, and prosthetics. Its ability to produce precise and biocompatible parts supports advancements in personalized healthcare solutions.
3. **Education**: FDM plays a crucial role in education by enabling students to learn about design and engineering through hands-on 3D printing projects. It promotes innovation and practical skill development in STEM disciplines.
4. **Science**: Researchers use FDM to prototype equipment for scientific experiments, build custom laboratory tools, and create models for visualization and testing purposes. It facilitates rapid iteration and customization in scientific endeavors.
5. **Automotive**: Automotive manufacturers employ FDM for prototyping vehicle components, tooling for assembly lines, and customized parts. It speeds up the design validation process and enhances efficiency in automotive engineering.
6. **Consumer Electronics**: FDM is utilized in consumer electronics for designing and prototyping product enclosures, casings, and internal components. It enables rapid iteration and customization to meet evolving consumer demands.
7. **Robotics**: Robotics engineers leverage FDM to prototype robot parts, create lightweight and durable components, and customize robot designs for specific applications. It supports innovation and optimization in robotic systems.
8. **Aerospace**: In aerospace, FDM is used to manufacture lightweight parts, complex geometries, and prototypes of aircraft components. It contributes to cost reduction, faster production cycles, and weight savings in aerospace engineering.
9. **Architecture**: Architects utilize FDM for creating detailed architectural models, prototypes of building components, and intricate designs. It aids in visualizing concepts, testing structural integrity, and communicating design ideas effectively.
Each industry example demonstrates how FDM enhances innovation, accelerates product development, and addresses specific challenges through advanced manufacturing capabilities.
INDIAN AIR FORCE FIGHTER PLANES LIST.pdfjackson110191
These fighter aircraft have uses outside of traditional combat situations. They are essential in defending India's territorial integrity, averting dangers, and delivering aid to those in need during natural calamities. Additionally, the IAF improves its interoperability and fortifies international military alliances by working together and conducting joint exercises with other air forces.
Coordinate Systems in FME 101 - Webinar SlidesSafe Software
If you’ve ever had to analyze a map or GPS data, chances are you’ve encountered and even worked with coordinate systems. As historical data continually updates through GPS, understanding coordinate systems is increasingly crucial. However, not everyone knows why they exist or how to effectively use them for data-driven insights.
During this webinar, you’ll learn exactly what coordinate systems are and how you can use FME to maintain and transform your data’s coordinate systems in an easy-to-digest way, accurately representing the geographical space that it exists within. During this webinar, you will have the chance to:
- Enhance Your Understanding: Gain a clear overview of what coordinate systems are and their value
- Learn Practical Applications: Why we need datams and projections, plus units between coordinate systems
- Maximize with FME: Understand how FME handles coordinate systems, including a brief summary of the 3 main reprojectors
- Custom Coordinate Systems: Learn how to work with FME and coordinate systems beyond what is natively supported
- Look Ahead: Gain insights into where FME is headed with coordinate systems in the future
Don’t miss the opportunity to improve the value you receive from your coordinate system data, ultimately allowing you to streamline your data analysis and maximize your time. See you there!
How RPA Help in the Transportation and Logistics Industry.pptxSynapseIndia
Revolutionize your transportation processes with our cutting-edge RPA software. Automate repetitive tasks, reduce costs, and enhance efficiency in the logistics sector with our advanced solutions.
Are you interested in dipping your toes in the cloud native observability waters, but as an engineer you are not sure where to get started with tracing problems through your microservices and application landscapes on Kubernetes? Then this is the session for you, where we take you on your first steps in an active open-source project that offers a buffet of languages, challenges, and opportunities for getting started with telemetry data.
The project is called openTelemetry, but before diving into the specifics, we’ll start with de-mystifying key concepts and terms such as observability, telemetry, instrumentation, cardinality, percentile to lay a foundation. After understanding the nuts and bolts of observability and distributed traces, we’ll explore the openTelemetry community; its Special Interest Groups (SIGs), repositories, and how to become not only an end-user, but possibly a contributor.We will wrap up with an overview of the components in this project, such as the Collector, the OpenTelemetry protocol (OTLP), its APIs, and its SDKs.
Attendees will leave with an understanding of key observability concepts, become grounded in distributed tracing terminology, be aware of the components of openTelemetry, and know how to take their first steps to an open-source contribution!
Key Takeaways: Open source, vendor neutral instrumentation is an exciting new reality as the industry standardizes on openTelemetry for observability. OpenTelemetry is on a mission to enable effective observability by making high-quality, portable telemetry ubiquitous. The world of observability and monitoring today has a steep learning curve and in order to achieve ubiquity, the project would benefit from growing our contributor community.
The DealBook is our annual overview of the Ukrainian tech investment industry. This edition comprehensively covers the full year 2023 and the first deals of 2024.
1. CONTENTdm Interoperability -- Leveraging resources; repurposing collections ALA Annual New Orleans, LA June 23 rd , Friday, 9 am to noon Claire Cocco , Product Manager Geri Ingram , Customer Service Specialist DiMeMa, Inc.
2. Agenda Part 1 9:00 to 10:15 Mainstream digital objects into existing workflows Importing from legacy systems Exporting Example of collaborative development for interoperability METS transform (courtesy of CDL) [BREAK 10:15 TO 10:30]
3. Agenda Part 2 10:30 to 11:30 Customizing and integrating your CONTENTdm site Web templates Custom Queries and Results Configuration files
4. Agenda Part 3 11:30 to Noon Handling Finding Aids Importing EAD files into CONTENTdm
5. Setting the context: fully engaged in digital library transformation Library services and collections expanding to encompass all Traditional to digital Licensed Reformatted Sharing Preserving
6. Leveraging resources Staff time and skills throughout the organization and/or consortium Existing metadata in some form Existing digital collections (images and transcripts)
7. Why? For better customer service In order to mainstream your processing and amplify your efforts. Your digital collections should ultimately be mainstreamed into regular workflows, similar to the ones used for other materials (whether that’s done centrally or in a distributed fashion). This includes selection, technical processing (cataloging, organizing, importing), integration with site vis-à-vis presentation and archiving.
8. Mainstreaming processing of digital formats (Part 1 of 3) Importing from other systems to CONTENTdm Exporting from CONTENTdm Example of collaborative development for interoperability CONTENTdm Standard Export METS transform for import
9. I . Importing from other systems to CONTENTdm Metadata only When records describe items that are not yet scanned Replace “null” files at later time Metadata AND their digital files
10. From an OPAC or other database system When you have… Individual image files cataloged already And can export from an OPAC or other dbms Or where you have compound digital objects ready for migration
11. Migration steps: Prepare the collection and the import files Cross-walk metadata to Dublin Core Configure the CONTENTdm collection fields Export and prep data in a tab-delimited ASCII file Import the file to CONTENTdm
12. Data prep: Common problems in tab delimited data files Extra data in columns or rows Extra tabs at end of line Extra CRs at end of file (Should only be 1 CR) Carriage return in metadata, tab in metadata Files must exist 0 versus O Error may occur in previous record, check few rows before and after error File names are required, not full pathnames
13. Data prep: Troubleshooting with Excel Use Microsoft Excel to open the file and view data Each row should be an item with last column as filename Work with small batches to find errors – keep adding items until record with error is found Use Excel’s “CLEAN” function to remove invisible characters Import images from directory without using tab delimited file Checks for any type of imaging errors
14. Demo : MARC to DC Export MARC records to tab-delimited text file (using ILS or MarcEdit) Format and clean up the text file to conform to your CONTENTdm Collection schema Import the file (with or without images) to the Collection
15. Importing compound objects For documents, postcards, monographs and picture cubes Can do singly or in batch Much easier to start with singles, then set up for batch when process is smooth
16. Migrate compound objects from another database system Where you have many compound digital objects to migrate Prepare the collection and the import files Cross-walk metadata to Dublin Core Configure the CONTENTdm collection fields Configure folders for scans and transcripts (if appropriate) Choose an import method based on your data structure Create tab-delimited ASCII file(s) appropriate to the method Import the files to CONTENTdm in batches
17. Multiple compound object wizard Documented in online tutorial Today’s demo described in handout Four import methods for multiple object loading Compound object (same as single, but upload batched) Directory Structure (most flexible and efficient) Object List (useful when NO page-level metadata) Job List Time allowing, demonstrate three different object types using 3 of 4 methods
18. Choose a multiple compound import method based on your data YES YES * YES Monograph YES * YES YES Documents * YES YES YES Postcards Object List (No page-level metadata) Directory Structure Compound Object * Will demo
19. Do you have page-level metadata for the compound objects ? Are your scan files separated into compound object directories? Create compound object directories for EACH compound object. No Yes DIRECTORY STRUCTURE Yes Do you have one tab-delimited text file containing ALL the objects? Are they all the same type of compound object ? Break up into batches by type No No OBJECT LIST Yes Do you have tab- delimited text files for EACH compound object? . DIRECTORY STRUCTURE . Create text file listing all compound objects and object metadata or create a text file for each compound object. No Yes No Yes
20. Every one of the four CONTENTdm compound object importing methods Requires object -level metadata Requires preparation File–naming, keeping sort order in mind Each object has own directory for scans May use tab-delimited text file(s) Accommodates transcripts
21. A word about descriptive page-level metadata Supported by some but not all 4 import methods NOT supported by Object List At page-level Title is only field required Technical metadata, can be generated by Template creator
22. More on transcripts Typescripts and transcripts Requires a field designated as the data type “Full Text Search” Inserted into the metadata field of the scanned page During import Through use of .txt file found, or By Template Creator If OCR Extension in use Or by “Directory Import” as with early versions of CONTENTdm Transcripts and typescripts are supported by all four methods (i.e., not considered “metadata” for purposes of this discussion)
23. Demo: Import Multiple Compound Objects Monograph using Compound Object method Postcards using Object List method Documents using Directory Structure method
24. II. Exporting from CONTENTdm To ascii tab-delimited with field headers To xml: Standard Dublin Core —only DC Custom—all fields, including local but not structure CDM Standard—all fields, including structure
25. III. Examples of collaboration for interoperability Web integration through search engines, RSS OAI harvesting Enable at collection or server level Choose to suppress <pagedata> or not WorldCat registration Open WorldCat integration
26. CONTENTdm and a new METS transform Info available on USC in July Code at SourceForge Windows-oriented
28. What is/are METS? Why is/are METS good? What is 7train? How do I use 7train? What do I get from 7train? How do I get 7train?
29. What is/are METS? METS (Metadata Encoding and Transmission Standard) is an XML-based standard for encoding metadata to describe objects (digital or otherwise) within a digital library. See http://www.loc.gov/standards/mets/ for more information
30. What is/are METS? METS metsHdr structMap dmdSec amdSec fileSec behaviorSec METS metsHdr structMap dmdSec amdSec fileSec behaviorSec Yellow elements/tags are required; all others are optional Metadata for the management of the object: technical details, object history, etc. Description of the structure of the object, i.e. how the files fit together What to do with the object: machine actionable instructions A list of files that make up the object Descriptive metadata - title, author, subjects, etc. Metadata about this particular METS - encoder, contact info, etc.
31. Why METS? To be able to add your objects to other collections and increase the visibility your institution's assets.
32. What is 7train? 7train is an XSL-based tool for converting XML documents - in this case CONTENTdm exports describing objects managed in the CONTENTdm system - into METS objects suitable for submission to a digital library system, such as the California Digital Library's Online Archive of California. 7train is a platform-independent, standalone tool that was designed to work on any system and to be simple to use.
33. How does 7train work? It is as easy as dragging your CONTENTdm XML export file onto an executable file.
37. References & Links 7train Home: http://seventrain.sourceforge.net 7train Download: http://seventrain.sourceforge.net/7train_download.html CONTENTdm: http://www.dimema.com METS: http://www.loc.gov/standards/mets/ XSL: http://www.w3.org/Style/XSL/ The California Digital Library: http://www.cdlib.org The Online Archive of California: http://www.oac.cdlib.org
38. Interoperability Librarians, Archivists… For Library Users OPEN WORLDCAT OAI MARC RECORDS OAI Web WorldCat Regional Union Catalog Other digital archives OAI OAI XML DC DC CONTENTdm Existing Libraries 10K/50K/ Unlimited Objects New Libraries Other CONTENTdm sites CONTENTdm Multi-Site Server OPACS
39. BREAK—15 minutes This concludes Part 1 To come after the break: Part 2 Customization Part 3 Finding Aids
40. Customizing and integrating your CONTENTdm site (Part 2 of 3) Web templates Custom Queries and Results Configuration files
41. CONTENTdm Web Templates Customizable for integration Designed to support broad range of users Small to large organizations Beginners to experts Use out of the box with minimal customization Basic customization requires minimal HTML skills Fully customize including advanced extensions Based on a PHP API ( Hypertext Preprocessor and Application Program Interface)
42. Basic Customizations Minimal skills needed Easy to make changes Global include files Variables Recommend all organizations do basic customizations Header (name/logo), contact e-mail address, colors, about page, home page http://www.contentdm.com/help4/custom/templates.html
43. Getting Started Access to Web server docs directory HTML editor or text editor Design plan Logo or other graphics Backup copy of original files
44. Customization Demo http://sr.contentdmdemo.com Files located in /cdm4 directory /includes/global_header.php /client/LOC_global.php /client/STY_global_style.php about.php browse.php results.php New logo saved in /cdm4/images/
45. Advanced Customizations Experience with HTML, PHP, and JavaScript needed Customize looks for each collection University of Nevada, Reno Web Template extensions E-commerce (University of Utah, Oregon State University) Comment forms (SENYLRC, Enoch Pratt Free Library, OSU) Custom metadata display (University of Oregon) QuickTime video (Williams College) http://www.contentdm.com/customers/index.html
46. Examples of Advanced Customizations University of Nevada, Reno http://imageserver.library.unr.edu/ University of Utah http://www.lib.utah.edu/digital/bodmer/ Oregon State University http://digitalcollections.library.oregonstate.edu/cdm4/client/bracero/ SENYLRC http://www.hrvh.org/ Enoch Pratt Free Library http://www.mdch.org/ Williams College http://contentdm.williams.edu/
47. Customizations Tips Always make a backup! Be aware of encoding (UTF-8 vs. ASCII) See what other users are doing Share, borrow, and copy ideas and code http://www.contentdm.com/customers/index.html Listserv Document changes Document which files are edited and what code changes are made to ease upgrading to newer versions
48. Custom Queries and Results (CQR) Create predefined, custom queries Virtual collections Guide users to specific results Integrate with other sites Multiple options Simple hyperlink, drop-down list, index box, text box, browse Easy to use Wizard generates code to copy and paste into Web pages Documentation http://www.contentdm.com/help4/custom/cqr.html http://www.contentdm.com/USC/tutorials/cqr.pdf
49. CQR DEMO Generate code using CQR Copy and paste into Web pages May need to change path Customize as desired
50. Configuration Files Customizable files that reside on the server Stop words Full text field stop words – fullstop.txt Automatic hyperlink stop words – stopwords.txt http://www.contentdm.com/help4/custom/stopwords.html Image viewer Customize how images are displayed – imageconf.txt For all collections or per collection http://www.contentdm.com/help4/custom/zoompan.html
51. Imageconf.txt Demo Located in the /conf directory on the CONTENTdm server Can change globally or for individual collections If you wish to change the zoom and pan default settings for a particular collection, copy the imageconf.txt file from the Server/conf directory to the index/etc directory of the collection(s) you wish to modify. Make a backup copy!
52. Introduction to Finding Aids How many of you have them? Are they digital documents or paper? If digital, are they XML? Basic: create documents, monographs, and use http protocol to link XML: use EAD DTD, and style sheet to display
54. Current EAD Support Import of EAD files Automatic text extraction from EAD files when: The file extension of the EAD is .xml. The file includes a header record beginning with DOCTYPE ead. The collection has a full text search field. The full text search field is empty when the item is added to the collection. Up to 128,000 characters extracted from the following fields and placed in the full text search field titleproper, title, unititle, persname, famname, corpname, genreform
55. Current EAD Support Display determined by style sheet XSLT CSS Client side parsing Affected by Web browser
57. EAD Demo Configure Full Text Search field Store DTD and style sheet on server Edit path to DTD and XSLT in EAD files Import (single or batch) Add metadata Custom thumbnail if desired Upload, approve, index
58. Custom EAD Extension Example by Oregon State University Terry Reese, [email_address] Customized Web templates Client side or server side parsing Integrates display in templates VBScript for extracting metadata from EAD to tab-delimited text file www.contentdm.com/USC/templates/index.asp
60. Announcing new exposure for your CONTENTdm Collections Collection of Collections http://collections.contentdmdemo.com/ (also featured at contentdm.com/customers) Harvesting metadata from Collection sites at: http://primarysources.contentdmdemo.com Uses CONTENTdm Multi-site server