Sunday, November 29, 2009

Monday, November 23, 2009

Friday, November 20, 2009

Muddiest Points

I have no muddiest points for this week.

Friday, November 13, 2009

Readings

Digital Libraries: Challenges and Influential Work
-Search and discovery is always difficult
-There is a difference between providing access to discrete sets of digital collections and providing digital library services - But what is that difference? I don't feel it was really addressed.
-There were early federal research projects that led to the digital library of today.
-This research led to many new developments, such as digital copies of journals and Google.
-There have also been digital library technologies that have come from non-federally funded projects, such as Yahoo, D-space and XML.

Dewey meets Turing: Librarians, Computer Scientists and the Digital Libraries Initiative.
-Digital Libraries Initiative brought librarians and computer scientists together.
-Analogy of the two disciplines being brought together and then having to deal with an undisciplined teenager (a.k.a. the World Wide Web).

Institutional Repositories: Essential Infrastructure for Scholarship in the Digital Age
Developments
-Repositories are now more affordable as it costs less for digital storage.
-Many universities look to MIT and their DSpace program as a model.
-Lynch thinks repositories should have the work of students and faculty.
-He stresses that there is a difference between scholarly publishing and scholarly communication.
3 Main Cautions about Institutional Repositories
1. I.R.s that try to exert control over faculty and student work will probably fail.
2. Don't put too many policy restraints on the I.R.
3. Institutions shouldn't create an I.R. hastily just because it's popular. They need to realize the commitment it takes.

Muddiest point

There is no muddiest point for this week.

Friday, November 6, 2009

Week 10 Readings

Web Search Engines - This was an interesting article, and I learned some new stuff about web search engines. Like the fact that crawlers need to deal with all sorts of issues - speed, politeness, excluded content, duplicate content and spam rejection. The main thing I took from the 2nd part was the fact that the web's vocabulary numbers in the hundreds of millions. This is due to the different languages, the new acronyms and the fact that people make up words all the time.

Current developments and future trends for the OAI protocol for metadata harvesting - I had never heard of the OAI (Open Archives Initiative) Protocol before this article. It seems to be a great idea though. As the article puts it, this initiative was started as a way to federate the access to different online archives. The article focused on three archives in particular that are still in development:
1. Open Language Archives Community
2. Sheet Music Consortium
3. National Science Digital Library
Each of these archives is unique in it's scope and mission and they all have their own problems to deal with when it comes to open access. The two main issues that affect the OAI itself are that of completeness and discoverability. They want to make it easy for users to navigate between the different repositories and they also want as much information available to the user that they can.

The Deep Web: Surfacing Hidden Value - I admit I found this to be the most interesting of the three readings. This article stated that most search engines such as Google and Yahoo are only searching the surface of the information available on the web. Apparently there is 7,500 terabytes of information in the Deep Web and there is one service that will search through all of it - that is Bright Planet. I still don't know what to think of this assertion. Would searching the deep web make the results more relevant to the user, or will it simply dilute the rest of the information on the surface?