[how to] Suitable database software for small business |
- Suitable database software for small business
- Buffered operations with MongoDB
- oracle blocking sessions
- Recommendation for RDBMS for large table
- Expanding a dataset based on a weight field
- Example of loading into RDS from S3 using SQL*Loader
- Mongo repair database "ERROR: error: exception cloning object in local.system.replset "
- Query sql-server db and insert results into PostgreSQL
- Are there any third party tools that can read and/or analyze SQL Profiler trace (.trc) files
- Select rows, where 3 columns have the same values
- Is there a DB that will allow the same field name with different cases?
- What's the optimal server configuration for fastest inserting in a MySql database in this scenario?
- MongoDB RAM requirements doubt
- Monthly backup of SQL server DB to PostgreSQL?
- Profiling PostgreSQL
- Relationship between professor and course
- Fixing wildcard expansion resulting in too many terms error (DRG-51030)
- pgsql slow query on table of 500 million rows
- Oracle Enterprise Manager Database Console - default login and password?
- Should I refresh query plan cache
- Push Data from MySQL to SQL Server
- Convert units of measurement
- export (mysqldump) databases from MySQL version 4.0?
- Where and how to store my weird datamart
- why would command text not execute at all?
- Avoiding performance hit from GROUP BY during FULLTEXT search?
- AWS performance of RDS with provisioned IOPS vs EC2
- sp_startpublication_snapshot Parameter(s)
- Pattern matching with LIKE, SIMILAR TO or regular expressions in PostgreSQL
Suitable database software for small business Posted: 24 Jul 2013 08:45 PM PDT First of all, I apologise if this is in the wrong area. I'm looking for a suitable database program for a small engineering business, which needs to log various jobs. Currently, we're just using an Excel spreadsheet shared over the cloud, however this has problems with conflicts when two people edit the file at the same time. What we want is something that can operate over the internet (so that multiple computers can access it from different locations), allowing multiple editors simultaneously. It also needs to be capable of conditional formatting. A screenshot example of what we currently have is here From my research I have a vague inkling that SQL would be a good base, but I'm really looking for a program that requires absolutely no background knowledge of how the language works. i.e something that isn't going to take a significant amount of time to learn to use. If there's info that I've not included, please let me know. Thanks in advance. |
Buffered operations with MongoDB Posted: 24 Jul 2013 08:22 PM PDT I have a server with 8GB of disk space running MongoDB. It collects 250,000 documents a day (spread across three collections), so the 8GB is only enough to last it maybe a week or so before I have to push the documents periodically to a server with more space. Unfortunately, the machine isn't online 95% of the time and connecting it to a dedicated MongoDB server with more disk space isn't an option either. My current approach is to connect to the machine every few days, transfer all of the documents, and wipe the database, but that seems a little dangerous and hacky. I looked into replication, but that will still have the disk space problem and from what I can tell. Sharding won't work either, as I'd like to maintain a complete duplicate of the database on at least the computer that I store the full database on. Does MongoDB have any features that I overlooked that can work for this scenario? Maybe some sort of read/write buffer that can be flushed? |
Posted: 24 Jul 2013 08:26 PM PDT I am trying to investigate some concurrency issues on Oracle 11 RAC. So far the most helpful tool I have is "blocking sessions" view of the Enterprise Manager, which tells me which query is blocking everyone else and what it is waiting for. However, to take advantage of that view, I need to catch the problem as it is happening. So I'm looking for a way to query oracle for historic data which would give me data similar to what "blocking sessions" screen has to offer in Enterprise Manager. |
Recommendation for RDBMS for large table Posted: 24 Jul 2013 05:22 PM PDT I have been asked for recommendation on the optimal RDBMS for a large amount of data. The largest table would contain 2TB of data. My own largest table is only 400GB (mysql, performs very well). The individual rows are going to be short, no blobs/etc, just a lookup table. Is PostgreSQL nowadays performing better than MySQL? Can I ask DBAs with tables in this order of magnitude for their experience? Oracle might be considered as well. The available hardware is probably a standard linux box with about 64G ram. |
Expanding a dataset based on a weight field Posted: 24 Jul 2013 07:44 PM PDT I have a table that contains records in an un-expanded form. Each record has an associated integer weight that essentially informs us how many times the record should be replicated in order to get the true population. Say, I have 3 records in my table. Once expanded, the dataset will be like this (note - I removed the I have tried to do this using cursors but it is taking a really long time to execute. Is there a clever way to do this really fast? Any predefined T-SQL stored procedure that achieves this? |
Example of loading into RDS from S3 using SQL*Loader Posted: 24 Jul 2013 04:25 PM PDT Does anyone have a working sample of loading data into an Amazon AWS RDS Oracle database using SQL*Loader where the source file is on S3? In principle it's clearly possible. The Amazon RDS Data Import Guide for Oracle documents using sqlldr. But the only sample there loads data from a local EC2 instance. I have worked with RDS and with S3 and with SQL*Loader before, but I haven't worked with all 3 together. One sample would be really useful. (Related advice would be welcome too... but a sample is all I'm really looking for.) |
Mongo repair database "ERROR: error: exception cloning object in local.system.replset " Posted: 24 Jul 2013 03:37 PM PDT I'm trying to repair a mongo database and am running into the following error early on: (replica set and host names have been sanitized for public posting) To attempt the repair I shut down the mongod and started it up with My mongod.conf consists of just two lines: The .conf normally did have a replSet declaration, but this was removed for the repair as I'm trying to run the repair on a secondary. (Per the docs repairs can only be done on primaries or standalones; so need to make it look like a standalone temporarily). Any idea what the problem could be? |
Query sql-server db and insert results into PostgreSQL Posted: 24 Jul 2013 01:10 PM PDT I have a query that returns data from sql-server and I need to be able to put the results into PostgreSQL. I'm using C# for the task. |
Are there any third party tools that can read and/or analyze SQL Profiler trace (.trc) files Posted: 24 Jul 2013 12:47 PM PDT I have a .trc file from a trace that a DBA did on one of my databases. I don't have the SQL profiler tool installed on my PC so I can't view the contents and analyze the trace log. Are there any third party tools that can read and/or analyze SQL Profiler trace (.trc) files? |
Select rows, where 3 columns have the same values Posted: 24 Jul 2013 11:52 AM PDT I have a structure like this: I now want to select all rows, where the values of the first three rows are equal, so the result of this query using the sample data would be: How can I achieve this, I'm kinda stuck at the early beginning... Thanks in advance! Alex |
Is there a DB that will allow the same field name with different cases? Posted: 24 Jul 2013 01:13 PM PDT Is there a DB that will allow a user to create a table with two fields, one calls |
What's the optimal server configuration for fastest inserting in a MySql database in this scenario? Posted: 24 Jul 2013 01:09 PM PDT I have a MyISAM table partitioned by range with 90 partitions representing the last 90 days. I expect this table to receive somewhere between 500 million to 1 billion records before i start rotating partitions (dropping old ones and creating new ones). What server configuration options would you put in your Any other tips would also be appreciated. EDIT: By 'fastest possible' i mean best possible insert speed and being sustained once it contains 100's of millions of rows. |
MongoDB RAM requirements doubt Posted: 24 Jul 2013 10:01 AM PDT I have a data file with 100 million documents. The number of documents is expected to always remain between 100 million and 200 million for atleast a year. I have 15 GB Ram. Currently the update operation takes 2 hours to load. Suppose I would like to increase the performance to require 40 minutes to load, will increasing RAM be required? How much RAM would I need to add in order to achieve that increase? Please do let me know if I must provide additional details. I have indexes for the updated fields as well as I have set --nojournaling option. Currently I have optimized the update process by using syncdelay 0 and then manually fsync-ing every 100K records. Then I implemented a system with only in-place updates rather than updates that add new fields or increase document size. Now I feel that I am at a point where increasing RAM might be the only way to go. |
Monthly backup of SQL server DB to PostgreSQL? Posted: 24 Jul 2013 06:13 PM PDT The company I'm working for has a SQL Server with read-only access. They use Crystal Reports hooked up to PostgreSQL for reporting. Is there any way to make it so I can move all the data from the MSSQL DB to PostgreSQL without user interaction? That seems to be the caveat to what I'm trying to do. They need to be able to run this report after I leave without having to interact with it during the process. Or am I looking at this the wrong way? Is there a way to save a "snapshot" of the SQL Server DB that can be manipulated in Crystal Reports? The ultimate goal is that since the DB is dynamic we need to be able to have a static DB at the end of the month that all the reports can be ran on without having to worry about it changing. |
Posted: 24 Jul 2013 06:24 PM PDT We have a system built with Python 2.7, Django, PostgreSQL 9.1 and PostGIS 1.5. In this system there is some extensive validation processes that are, well, intensive. Our machine: SO: CentOS 6 HW: 8 cores, 32GB RAM, 4Tb hard-drive; We, for development and testing, are using the default PostgreSQL configuration, what I think, is somewhat wrong. I want to take advantage of the full capacities that our hardware provides, and so far we didnt got a lot of traffic. When this is release to the public, there will be a lot of users concurrently. Our application has some queues and async strategies to handle the load, but, the main question is: how can I safely tune PostgreSQL, run some tests, and assure that the tuning was somewhat effective? |
Relationship between professor and course Posted: 24 Jul 2013 12:42 PM PDT I'm designing a database where I have a Course table and and Professor table. The business rules are : 1) A professor can teach many courses. (This is obvious). However, I get stuck in the second rule. 2) On a given semester, a course can be offered on different schedules, each of those offerings can be taught by a different professor. For example a typical freshman course on calculus has about 100 students every semester. The course is divided into 4 sections, each has 25 students and a different professor. I would like to treat each combination of (course,professor) as a unique entry.For example, (Calculus , Prof Joe) and (Calculus, Prof Jhon) are two different courses. How do I implement this? Thank you |
Fixing wildcard expansion resulting in too many terms error (DRG-51030) Posted: 24 Jul 2013 03:54 PM PDT How can I resolve the My index creation syntax is: And my query is: There are 15 million rows in that table, and more than 50000 rows match that query. I have set
Can anybody hep me how to solve this error? |
pgsql slow query on table of 500 million rows Posted: 24 Jul 2013 09:25 AM PDT I have a table that keeps measurements of latencies between nodes running MPI tasks in a large cluster. The table looks like this: Now after a large experiment I collected over 500 million rows of data. I find querying these data painfully slow, below is an example of a SELECT COUNT(*) I am running both the PgSQL server and client on the same machine, which has 4 Xeon E7-4870s (40 cores/80 threads in total) and 1 TB of RAM. The effect of Linux file caching is obvious: the first query took well over 12mins while the subsequent ones took about 1.5min. Is there anything I can do to make the query run faster, since 1.5min isn't exactly responsive. Thanks. |
Oracle Enterprise Manager Database Console - default login and password? Posted: 24 Jul 2013 02:12 PM PDT I've just installed an Oracle 11g Standard Edition Database. It ran successfully and everything, however when going to Oracle Enterprise Manager Im not really sure what my "username" would be. I entered a password during the installer, and I tried SYS and my password connecting as "Normal" and "SYSDBA". However, I'm not really sure what my username is meant to be. Is it just that I haven't set up a username/password for this yet? Thanks |
Should I refresh query plan cache Posted: 24 Jul 2013 10:10 AM PDT Please let me explain my problem and situation: I have a web application - MVC3, MSSQL Server 2005, LinqToSQL. It has been running great until one fine morning I pushed a lot of rows to a table that is heavily used and since then I was getting query timeouts. In order to fix the problem I run the Database Tuning Advisor and I added some Indexes and Statistics. I also created a maintenance plan to rebuild indexes daily. After those additions, the application has been behaving unstable; it would work fast for couple of hours then it would start timing out again. Next, life forced me to clean up the table in matter, and the amount of rows in it is even smaller now than it was before but the timeouts are still happening. So, I removed all indexes that I created and now the website is much more stable but from time to time I still see some timeouts. I've been trying to figure out how to fix those queries and when I profile it and paste the query directly into the SQL Management Studio it returns the results in 1 second, but when I run this query from my application, it's about 25 seconds. Then after it runs for the first time, next time it goes as fast as on the server! I started doing some research and it looks like when I played with all those indexes my query plans got messed up and now they are creating issues. My questions are :
|
Push Data from MySQL to SQL Server Posted: 24 Jul 2013 01:50 PM PDT I am not sure if my title is correct, but I will try my best to explain the current problem. Current Scenario:
I have completed this:
I have been reading and trying the following:
Problems:
Any detail explanation is greatly appreciated. Thanks Taken from my question in: http://stackoverflow.com/questions/17583924/push-data-from-mysql-to-mssql |
Posted: 24 Jul 2013 06:28 PM PDT Looking to calculate the most suitable unit of measurement for a list of substances where the substances are given in differing (but compatible) unit volumes. Unit Conversion TableThe unit conversion table stores various units and how those units relate: Sorting by the coefficient shows that the This table can be created in PostgreSQL using: There should be a foreign key from Substance TableThe Substance Table lists specific quantities of substances. For example: The table might resemble: ProblemHow would you create a query that finds a measurement to represent the sum of the substances using the fewest digits that has a whole number (and optionally real component)? For example, how would you return: But not: Because 112 has fewer real digits than 1.12 and 112 is smaller than 1120. Yet in certain situations using real digits is shorter -- such as 1.1 litres vs 110 centilitres. Mostly, I'm having troubles picking the correct unit based on the recursive relation. Source CodeSo far I have (obviously non-working): IdeasDoes this require using log10 to determine the number of digits? |
export (mysqldump) databases from MySQL version 4.0? Posted: 24 Jul 2013 09:37 AM PDT Goal: export the schema and data from a MySQL version 4.0.16 DB. Background: I was asked to migrate a database from MySQL version 4.0.16. The database was installed on a Windows XP machine, and I just could not find mysqldump tool on that machine. I then putty into a linux machine that has mysqldump. I ran the command, but got an error: I looked up mysql reference. INFORMATION_SCHEMA is something new to version 5 and above. This means the mysqldump on the linux box is incompatible with the production MySQL server which is on v4.0.16. I tried to download Mysql 4.0 server and install it on a windows machine so that I can get hold of a compatible version, but searching on Mysql website found nothing older than 5.0. (I also tried Mysql workbench to connect to this DB. Failed. "Unknown character set utf8") So how can I export the schema and data at all from this legacy mysql DB???... |
Where and how to store my weird datamart Posted: 24 Jul 2013 02:55 PM PDT I could really use some help here. Here's my situation. I'm building a datamart in SQL Server 2005, which will furnish a single report (for now). We have 26 clients (medical organizations) that will be using this report. Each client will have between 1 and 2,000 users accessing it at any given time (most likely to average in the 50's but it needs to scale). There will be 26 versions of the same rdl, one for each client. Each one will access its own respective client database. The interface for said report is SSRS 2012. This is our first 2012 report - the rest are still 2005, and the rdl will access a 2005 database. I am not using OLAP; the SSRS report runs SPs and views. We built a virtual server and installed SQL 2012 to host the report (both these virtual servers live on the same physical machine, in case that matters). Nothing else will be running on the SQL 2012 virtual server. These are the facts about the environment... Our system is not OLTP heavy - with one exception, which I'll describe below, it's all Read (except for the ETL of course). We have one client-facing database per client, 26 altogether. In these we store transactional data, rolled-up data, some report-ready flat tables and a ton of T-SQL code that crunches numbers when clients pull up reports in SSRS. I'll refer to these databases as "operational" because, for me, they will essentially function as ODS's. The operational databases are loaded by a horrific ETL process (clients have varying load schedules - usually monthly or weekly). I'll be building a mini ETL process (hopefully not so horrific) to populate the datamart from those operational databases. All of the datamart's dimensions are conforming, but due to HIPAA restrictions, some of them (like Physician and Patient) can't be stored in a central database, and neither can the fact table. So there will need to be 26 versions of the same fact and dimension tables, one for each client. There is a real-time component to our system. Physicians and nurses can enter transactional data via our site and all reports need to reflect the changes immediately. As for the datamart, they'll only affect the fact table. For what it's worth, this is why I decided not to use SSAS. I know diffs process really quickly but it just feels like too many moving parts. I plan to create a trickle-in fact table with a view that combines it with the main fact table. Again, I'll need 26 of these. A pared down version of my new ETL process will need to run upon each user edit. Here are my questions...
With all of these questions, I'm concerned about good design practice, but mostly about the performance of the report and the performance of the ETL that will need to run upon user edit. I hope all of this made sense. I would very much appreciate any feedback! EDIT: @Jon Seigel - There will be 26 versions of the same rdl, one for each client. Each one will access its own respective client database. EDIT: @JNK - I merged accounts and read the faq's. Hopefully I'm responding correctly now. |
why would command text not execute at all? Posted: 24 Jul 2013 05:38 PM PDT in my report i have:: i am using sql server profiler to see exactly what statement is being set. i have set only two filters:
yet after running the report, no statement gets intercepted. i suspect that because i am a beginner at SSRS, i am missing something crucial here. for what reason would commandtext not be executed at all? i did follow this question, to make sure that i am using sql profiler correctly, and indeed, i am: http://stackoverflow.com/questions/9107383/sql-server-profiler-capture-calls-to-your-databases-stored-procs-during-ssrs another bit of important information is although the chart shows no data: i actually am indeed showing data when i run the commandtext from ssms! |
Avoiding performance hit from GROUP BY during FULLTEXT search? Posted: 24 Jul 2013 10:37 AM PDT Is there any clever way to avoid the performance hit from using group by during fulltext search? In this example it's fetching the lowest post_id for unique topic_ids that match the text. With the group by to find the min, it's taking 600ms in a million row database, with about 50K rows examined. If I remove the MIN but leave the GROUP BY, it's the same slowness, so it's the GROUP hit. I suspect this is because it can only use one index, the fulltext ? Without the GROUP BY it's 1ms so this has to be filesort speed? (I've removed ORDER BY and everything else to isolate where the hit is) Thanks for any insight and ideas. (using MyISAM under mariadb if it matters) |
AWS performance of RDS with provisioned IOPS vs EC2 Posted: 24 Jul 2013 11:37 AM PDT Has anyone done a performance comparison of AWS RDS with the new provisioned IOPS vs EC2? I've found plenty of non-high IOPS RDS vs EC2 but nothing with the new high IOPS feature in RDS. |
sp_startpublication_snapshot Parameter(s) Posted: 24 Jul 2013 06:38 PM PDT I am creating a stored procedure that:
Here is my new brick wall... On this DB server, I have a 'shell' db that they will be running the SP from, that has a history table so I can keep track of who created/deleted databases using my SP's... The only parameter for sp_startpublication_snapshot is @publication... I can give it the publication name, but since I am not running it from the publishing database, how do I specify the publishing database? i.e.: the publication shows up as: but I am running the script from the database [WC_QACatalog] Any ideas about how to accomplish this? Thank you, Wes |
Pattern matching with LIKE, SIMILAR TO or regular expressions in PostgreSQL Posted: 24 Jul 2013 02:09 PM PDT I had to write a simple query where I go looking for people's name that start with a B or a D : I was wondering if there is a way to rewrite this to become more performant. So I can avoid |
You are subscribed to email updates from Recent Questions - Database Administrators Stack Exchange To stop receiving these emails, you may unsubscribe now. | Email delivery powered by Google |
Google Inc., 20 West Kinzie, Chicago IL USA 60610 |
Thanks for sharing great post with us about SQL. SSIS Postgresql Write
ReplyDeleteThis comment has been removed by the author.
ReplyDeleteThis comment has been removed by the author.
ReplyDelete