获取的数据是比较有规律的,因为差不多每秒要从对方服务器获取差不多600多条的数据,不知道是生成一个页面,然后把内容转化为数组好,还是把内容存到mysql中要好,
如果说存到数据库里面去,就要差不多每秒更新600多条记录.主要要考虑到以下几点1,数据大概有600多条100K;
2,差不多一秒或者二秒要从对方服务器来抓取数据;
3,抓取的数据是供很多会员来查询的.
4,要考虑到会员能够在最短时间获取最新的数据.抓取的内容是以下var $hq_str_sh600000=array("浦发银行,23.27,23.17,23.82,24.15,23.15,23.81,23.82,93155951,2207129971,6345,23.81,247300,23.80,2100,23.79,32000,23.78,500,23.77,10800,23.82,99710,23.83,91280,23.84,66160,23.85,23300,23.86,2009-05-04,15:03:31)";
var hq_str_sh600001="邯郸钢铁,4.21,4.18,4.40,4.42,4.20,4.40,4.41,58768722,256886428,145645,4.40,554200,4.39,517049,4.38,273600,4.37,600812,4.36,654425,4.41,978580,4.42,669000,4.43,622193,4.44,1714075,4.45,2009-05-04,15:03:31";
var hq_str_sh600002="";var hq_str_sh600003="ST东北高,0.00,3.87,0.00,0.00,0.00,0.00,0.00,0,0,0,0.00,0,0.00,0,0.00,0,0.00,0,0.00,0,0.00,0,0.00,0,0.00,0,0.00,0,0.00,2009-05-04,15:03:31";
var hq_str_sh600004="白云机场,8.82,8.82,9.05,9.06,8.80,9.04,9.05,16247215,145205420,2000,9.04,93189,9.03,76193,9.02,76422,9.01,94434,9.00,45388,9.05,20900,9.06,44300,9.07,78440,9.08,44211,9.09,2009-05-04,15:03:31";
var hq_str_sh600005="武钢股份,6.93,6.86,7.35,7.38,6.93,7.34,7.35,184641695,1338634442,376209,7.34,259300,7.33,1191415,7.32,580800,7.31,672656,7.30,5300,7.35,30500,7.36,681627,7.37,1460132,7.38,1113668,7.39,2009-05-04,15:03:31";
var hq_str_sh600006="东风汽车,4.33,4.30,4.38,4.40,4.27,4.37,4.38,20443835,88945823,206106,4.37,254224,4.36,245199,4.35,122800,4.34,229450,4.33,48900,4.38,290519,4.39,693803,4.40,96300,4.41,243350,4.42,2009-05-04,15:03:31";
var hq_str_sh600007="中国国贸,8.80,8.77,9.00,9.00,8.74,8.99,9.00,4723692,42179082,23900,8.99,43003,8.98,42400,8.97,138300,8.96,27900,8.95,139976,9.00,18500,9.01,18600,9.02,37700,9.03,8800,9.04,2009-05-04,15:03:31";
..........................

解决方案 »

  1.   

    1. 100K不算多,但你那600多条是指600多个页面吗? 如果是...小心被封锁
    2. 1~2秒,如果项目是相同的,只是为了更新数据的话...这个频率太高了(完全没有必要)
    3. 如果数据已经格式化的话~当然是直接存入数据库
    4. 数据没变化的话,数据库是不会重写记录的
      

  2.   

    每秒都抓取,过于频繁.
    可以考虑不要获取重复的数据.
    存文件和存数据库都可以.
      

  3.   

    每秒都抓取,过于频繁.
    可以考虑不要获取重复的数据.
    存文件和存数据库都可以.
      

  4.   

    不考虑 去重复的话
    写文件较快
      

  5.   

    数据库肯定是首选啊
    那么多数据,操作文件效率会很差的
      

  6.   

    因为我获取的是股票数据,股票数据当然是实时在更新的,我从对方服务器一次接受600支股票数据,然后存在一个页面中,其它会员引用这个页面,是不是要好些,否则就要将600支股票进行分析,然后循环更新到数据库中,因为更新速度很快,我怕影响到数据库,因为在更新的同时有好多会员在查询这些数据