Well there are many answers and my experiences have shown that it always depends on the use-cases.
What you can do to make inserts faster:
1) Bulking (if you want to make it fast, do bulking!) which reduces commits and commit time
2) Enough ITL on the block (INITRANS)
3) Depending on the size of a row: Blocksize, PCTFREE
4) No Indexing
5) No integrity checks (Foreign keys, not null constraints)
6) Buffer cache (if it's to small you'll early end up with I/O)
And the list goes on....
APPEND hint is also very useful on inserts because it's using DIRECT-PATH inserts which will append the data simply on the table. But be aware the unused space by the table will then not be filled up!
Gerald made an excellent exposition and I totally agree withg him.
From my personal experience I would recommend partitioning. But my experience is in the telecom world were we could easily make 20/30 million inserts per hour and some of these tables were partitioned by (day,hour) to improve insert speed, we had indexes and we could not use /*+append*/ because of backup restrictions.
If you need real insight or your specific case, may be you could post your full scenario.
Hi, do you have a column in the table to store the radius server id (i.ex server1, server2, ...)?
If so, is it indexed? Is it partitioned?
maybe you could consider partitioning the table according to the server. For instance
if table looks like this
c1, c2, radius_server_id, ...
You could use range or list partition by radious_server and the on each server, when you issue the insert statement you could specify the corresponding server id.
If your planning to rebuilt table and add partitions, i don't think hash would do the trick on this one...
I stand by adding a range or list partition using a column server_id and forcing each server to specify its own partition while doing the insert.
insert into abc partition (s2_server) values(1,2,'s2') ;
I'm not a fan of advertising, and so I will not be including any advertisements on OracleCommunity.net. However, managing this community does not come without cost! If you are willing to donate to help pay for the monthly community fees and domain services I accept Bitcoin and PayPal donations.
Companies growing continuously have to deal with large volumes of data on a regular basis. It becomes really difficult for these business organizations to manage this data. It eats up a lot of time and diverts the attention of the top level management from core-business tasks. In such cases, outsourcing proves out to be a better option. By outsourcing repetitive data entry tasks to professionals, companies cut costs, increase their revenue and improve the productivity of business. They can…
One of the most valuable assets of an organization, data lies at the heart of business. It impacts everything from customer service to email deliverability and ultimately revenue generation. If the data that you have is not accurate, it will have a direct impact on your company’s ability to meet…
Data Mining is the process of analyzing and extracting crucial information from a database and presenting the same in an appropriate way to help business cut cost and boost revenue. In simple words it is the process of drawing crucial and actionable insights, which is imperative for better decision making.
Data mining is crucial for effective management of your business. It is also used for verifying whether or not the strategies are going as planned so as to provide desired results.…
In order to focus on crucial business activities, organizations worldwide prefer outsourcing data entry services to reliable service providers. Though redundant but highly crucial, data entry still remains an important component of document management and requires vigorous quality checks. Outsourcing data entry tasks to experts ensures affordable and quality output while saving time and resources that can be…