So we at Robelle have installed the SAPdb enterprise-class open-source database and put it through some initial tests. SAPdb is designed for the business enterprise, with 24x7 uptime, scalability and high performance in mind. There are no limitations on database sizes or on the number of users. SAPdb is ACID compliant (fully supports ISO-SQL 92 Standards) and includes all RDBMS and enterprise features expected in an open DBMS such as Views, triggers, foreign keys, constraints of various kinds, stored procedures, versioning, hot backups, etc. ODBC and JDBC drivers are supported.
If you are thinking of using SAPdb as your migration target, you need to understand the datatypes available to you. They are similar to those in Oracle, SQL Server and other SQL databases. Read our explanation of the SAPdb datatypes.
SAPdb is designed for the business enterprise, with 24x7 uptime, scalability and high performance in mind. There are no limitations on database sizes or on the number of users. SAPdb is ACID compliant (fully supports ISO-SQL 92 Standards) and includes all RDBMS and enterprise features expected in an open DBMS such as Views, triggers, foreign keys, constraints of various kinds, stored procedures, versioning, hot backups, etc. ODBC and JDBC drivers are supported.
Click the link above for a full report.
I was looking for training on Microsoft Foundation Class (MFC). This is taught at many sites across North-America, but only a couple are found in the Montreal area. The courses at these sites are scheduled at most twice a year. Last year I signed up for one, just to have it cancelled at the last minute.
Most courses last 4-5 days and cost around $1500US. Of course, this excludes travel expenses. Since I did not want to travel, I decided to look at web-based training (a.k.a. e-learning). There were also a number of offerings. I decided to sign up with Digital Think (http://www.digitalthink.com).
They offered MFC Fundamentals I and II at $325US each or $575 for both. Each course should be completed in 15 hours.
After signing up, I was granted access to the course material for 6 months. Registration procedure was easy, all online if you want to use a credit card. You can register via fax or phone if you want to use a different means of payment. If you pay by credit card, you have immediate access.
These courses required that I had Visual C++ installed on my PC in order to do the exercises.
There were functions allowing interaction with other people taking the same course. I could post questions on a discussion list where everyone else could read and reply, if they wished. Sounds like a nice feature but, from what I could tell, it was not heavily used. Most messages on there dated back to 1999 or so. I didn't feel the need to post anything there either.
I could also have one-on-one chat sessions with other students. Again, it didn't seem like it was used. Of course, people are taking the course at their own pace. They have 6 months to go over 15-hours worth of material. I guess not many people were actually "on the same page" at the same time, which made it hard to find a match.
At regular intervals, there were interactive quizzes. These were short multiple-choice questionnaires. They were graded automatically as soon as I had done them. Along with the results, they provided a short explanation as to why each answer was wrong or right. There were also exercises where I was given a short problem to solve, typically working on a Visual C++ project that was enhanced as I went along. In most cases, they asked that I send pieces of code to the tutor to show I had done it, and to get graded.
Some features were shown as simulations. Screenshots of a Visual C++ session were displayed with step-by-step instructions. There were also "mouseover" graphics that displayed pieces of code. I simply moved the mouse over individual statements to see a short description. The mouseover graphics did not always display properly when scrolling with the wheel. I had to use the scroll bar to get it perfect.
Parts in the Orientation module were not up-to-date, actually dating back 2-3 years. Course material was also available as PDF files.
At the end of each course, I had to fill out a survey. I could also request a completion certificate (PDF) to print for my records.
Overall, the e-learning experience worked well but it was not that different from using self-paced training books such as SAMS's "Teach Yourself" series. A good book probably covers a lot more material. Most books don't provide quizzes and definitely not a tutor. I think the material covered in these courses was fairly basic for my needs. The tutor was also of limited use for me. My guess is the equivalent instructor-led training classes would have been a bit more complete. However, I wouldn't have known until the end of the week i.e. too late.
It was a lot cheaper than signing up for instructor-led classes but, for the same price, I could have bought a few good books and gone through them in a week or two.
Recently a friend needed to work with some data in mySQL to test a Web Application. I figured that I could provide some test data from one of our IMAGE databases and use our STExport program (part of the Suprtool product) to move the data over.
mySQL is an Open Source database that is commonly used as the backend database server for many Web applications for a huge number of Linux and Unix platforms as well as Windows machines.
In order to get started we decided to build a mySQL database that looked like an Image database, building a simple Master dataset and a single detail dataset. The byte type fields in Image were created as char fields. The I2 or J2 fields were created as int fields. The database resided on a small Windows 98 laptop system.
Having done this, we extracted data from the 3000 database and used default STExport settings to output a file that was comma delimited, with each field enclosed in quotes.
In investigating how to import data into mySQL, we first tried the mySQLImport program, but it didn't seem as robust and we could not figure out how to tell it what delimiters to use.
In looking at the documentation, we thought that the LOAD_FILE command might work, but further investigation showed that this command opens the file and returns the contents as a string. This feature is only used by Text and Blob columns of mySQL.
We finally had success with the LOAD_DATA statement after reading the documentation on the mySQL Website at:
http://www.mysql.com/doc/L/O/LOAD_DATA.htmlWe quickly tried importing the data with the default STExport settings. However, when we looked at the data, some records were not imported correctly. The records seemed to have the data offset by one field. We found the problem to be records with either quotes or commas in the data. Since these were also our delimiters and seperators we changed the Suprtool and STExport commands to the following:
get m-customer out mcust,link xeq export in mcust export quote none export delim "?" export out mcustexp export exitWe got rid of the quotes surrounding each field by using the Quote None statement and changed the delimiter to a character that we knew did not exist in our data. In our case we chose the Question Mark.
Importing the data into our table then became a single simple command entered from within Mysql Monitor:
load data infile 'm_customer_file.txt' into table m_customer fields terminated by '?';Once we had the basics down for importing the datasets into the mySQL tables, we could then import the entire database into mySQL.
We were able to check the data using the mySQL admin tools, but for more flexible access we installed some PHP scripts for working with mySQL databases.
Overall, we found it relatively easy with STExport to duplicate our production IMAGE database structure and contents in mySQL. Of course, we did do a little studying in two books:
When you use the List command on a file, you can use the $DBL option:
/listq $dbl myfile lpIf you want to double-space your file (i.e., insert a blank line between all existing lines), use the Divide All command. The default is divide each line after the last nonblank character.
Qedit for Windows does not have a function for double spacing, although with the Qedit
Scripting Language you could write a script to copy your file into a new workspace,
insert a blank line after each existing line, and then print the file.
If you want your output for List Standard to be double-spaced, one sneaky but effective way to do this is to append a string of blanks to the extracted fields.
Insert enough blanks to force each record to wrap around to a second line:
>extract field1,field2,field3,field4 >extract " "
Another approach would be to extract a CR/LF (^13,^10) at the end of record, although this would probably only work on PC style printers, not line printers.
This is a timely article, since LDAP is now available on the 3000.
But you don't need a huge company, just a computer and a part-time person. So you don't need to have a 100-person company to develop that idea. You can do it in your spare time, you can really work on ideas and see if they take off -- rather than trying to raise tons of money, millions of dollars for an idea that may or may not work. And once you have the product and people are using it, it's very easy to raise investment.
!job setclock,manager.sys !comment !comment Daylight Savings adjustments ! !showtime ! !IF HPMONTH = 10 AND HPDATE > 24 THEN ! ECHO Back to Standard Time ! SETCLOCK TIMEZONE = W8:00 !ENDIF !IF HPMONTH = 4 AND HPDATE < 8 THEN ! ECHO Ahead to Daylight Time ! SETCLOCK TIMEZONE = W7:00 ! !ENDIF !showtime !eoj
It is very useful to have a small refrigerator in your room, for storing cold drinks, snacks, and something for breakfast. But many hotel rooms do not come with a refrigerator. Did you know that you can call the hotel desk and ask for a fridge in your room? Most hotels have them available, due to the requirement of diabetics to keep their insulin cooled, and will add one to your room if you just ask!
I recently attended a one-day course on "Presenting Data and Information" given by Edward Tufte, author of the incredible books "The Visual Display of Quantitative Information", "Envisioning Information", and "Visual Explanations". Tufte is Professor Emeritus at Yale University, where he taught courses in statistical evidence, information design, and interface design. Read about this excellent tutorial.
Unfortunately, this "millenium edition" contains a pre-preface which assures readers that what follows is a work of fiction. Dewdney's account is so completely matter-of-fact -- he is, after all, a college professor writing about students working on a class project -- that it's much more fun to pretend to read it as a documentary.
This is one book I've never been willing to lend out. I recommend it most highly. Amazon link. - Bruce Toback
Sounds like a good cause for all our UK customers ...
p.s. His editor will soon be renamed as 'RTF Edit control' if you're interested.
Paul Gobes, Robelle Technical Support.