2007-06-21, 05:13 AM
sub Wrote:Its fine, post the name.
Dam you be fast! ctpvr
This is without a doubt the most responsive community I've seen.
2007-06-21, 05:13 AM
sub Wrote:Its fine, post the name. Dam you be fast! ctpvr This is without a doubt the most responsive community I've seen.
2007-06-21, 05:17 AM
sub Wrote:If you find away to extract that data into the common xmltv file format, then you'll be able to use it in GB-PVR Where could I get a sample of a common xmltv format? The data is in a access file. I should be able to export using access right?
2007-06-21, 05:26 AM
cpa Wrote:They were in violation of the zap2it license for a long time, but I remember they went legit late last year. They dont have digital channel listings at this stage though.fender Wrote:I would gamble that any PVR app you bought for three dollars is one of the culprits that is violating Zap2It's non-commercial licensing mentioned on their website.I could not tell you. How would I find out if that is the case?
2007-06-21, 05:28 AM
cpa Wrote:Where could I get a sample of a common xmltv format? The data is in a access file. I should be able to export using access right?Its a pretty simple file format: http://xmltv.org/wiki/xmltvfileformat.html
2007-06-21, 05:30 AM
sub Wrote:Its a pretty simple file format: http://xmltv.org/wiki/xmltvfileformat.html Much thanks, I'll be baaack if I getting to work :confused:
2007-06-21, 06:02 AM
I'd like to note, if this hasn't been discussed already, that scraping web sources is problematic if program descriptions (and you care about them) aren't displayed on the same page as the show grid (Titan: no, zap2it: no, TV Guide, no). My typical lineup of less than 130 channels for 14 days gives me a little over 31,000 programs. If you get a grid that displays all your channels for 6 hours, you'll need 56 HTTP requests to get the grid and another 31,000 - one for every program to get its description.
If you can get each description in exactly one second, which I don't think is realistic, you'll be looking at 8.5 hours minimum and enough requests to get you banned from the server and possibly threatened legally. Even if it can be done in 1/3rd the time, that's still nearly three hours of hammering that webserver (even faster) which is even more likely to get you yelled at. I for one don't generally care about descriptions enough for this to be a problem, but for those that do care, it's not going to happen. As for schedule scraping, I got half way through writing a perl scraper for zap2it's current public interface in just over an hour, outputting pseudo-XMLTV compatible xml. Don't worry too hard, that will come out soon from probably a dozen different people or more, it's just grunt work. I don't see getting more than that anytime soon though. This well and truly sucks.
2007-06-21, 06:12 AM
pwtenny Wrote:I'd like to note, if this hasn't been discussed already, that scraping web sources is problematic if program descriptions (and you care about them) aren't displayed on the same page as the show grid (Titan: no, zap2it: no, TV Guide, no). My typical lineup of less than 130 channels for 14 days gives me a little over 31,000 programs. If you get a grid that displays all your channels for 6 hours, you'll need 56 HTTP requests to get the grid and another 31,000 - one for every program to get its description.This is pretty much what I said a few pages back. I even used 30,000 as an example of numbers of web pages hits and how I got banned for doing exactly this. After the initial load you could drop back to scrapping just the next 36hrs or so of listings, which drops the number down quite a bit - but its still too high. This is why I was suggesting other options for propagating the info in chunks via usenet or p2p, though this probably steps over a legal line (though with all the people willing to share movies and music online, sharing a few TV listings might not be a concern to many).
2007-06-21, 06:58 AM
madcat Wrote:let me get this straight - the effort here is to write some kind of parser that will parse web page(s) into XML info of EPG? this two points are the main problems of most xmltv grabbers. so far i wrote three grabbers (tv_grab_ch, tv_grab_ch_bluewin, tv_grab_ch_search). two of them got blocked by the source provider after a couple of months. tv_grab_ch_search (http://tv.search.ch) seems to tolerate it (as e.g. tvtoday.de does since years). a change of the layout usually means a change of the grabber. as most authors of the grabbers use them for themselves, things like that are spotted and taken care of fast.
AMD 1600mHz, 512 M RAM, 20+160 GB drive, Hauppauge PVR-350 + PVR-150 + PVR-150MCE, Win XP Pro SP3, using software decoding.
There is a theory which states that if ever anyone discovers exactly what the Universe is for and why it is here, it will instantly disappear and be replaced by something even more bizarre and inexplicable. There is another theory which states that this has already happened. (The Restaurant at the End of the Universe)
2007-06-21, 07:22 AM
(This post was last modified: 2007-06-21, 08:09 AM by dragonfiregum.)
I assume a xmltv grabber cant draws data from the cable provider like a cable box loads its own epg?
Or is it possible to draw straight from your cable provider? i mean my cable box knows the schedule somehow right? I understand this is probably encrypted... but if you are paying for cable (which you obviously should be if you watching those channels ) that data comes with your subscription to cable. Pulling it for tv watching/personal use should be legal as long as your not redistributing it. Is there any ways to draw data from the line then? If not maybe a shot in the dark but yahoo seems intrested in tivo. Maybe they would offer a monthly fee or something for data. But i have no idea how to contact them as major companys like that dont seem to like to include contact info anywhere.
---------------------------------------------------
Microsoft Windows XP Home Edition Pentium D 805 2.66ghz Dual Core 2GB Ram Nvidia Geforce 6600 GT 256mb PCI-E (x2sli) Audigy ZS 2 Platinum Sound card ATI Theater 550 PRO w/Remote 3 320gig Harddrives(and 400+external on other 2 pc's which are setup)
2007-06-21, 10:24 AM
sub Wrote:This is pretty much what I said a few pages back. I even used 30,000 as an example of numbers of web pages hits and how I got banned for doing exactly this. After the initial load you could drop back to scrapping just the next 36hrs or so of listings, which drops the number down quite a bit - but its still too high. This is why I was suggesting other options for propagating the info in chunks via usenet or p2p, though this probably steps over a legal line (though with all the people willing to share movies and music online, sharing a few TV listings might not be a concern to many). Yeah, I got my number wrong though. Quote:Downloading EPG data from Tribune... Since the only way to get data into GBPVR now is through XMLTV files, I have a few questions relating to that. XMLTV's format seems to consist of the two following things: Quote: <channel id="I10021.labs.zap2it.com">For a channel, and.. Quote: <programme start="20070616013000 " stop="20070616020000 " channel="I30017.labs.zap2it.com">..for a program. Which elements from this are required for GBPVR to be able to parse the file? There isn't much you can get from scraping beyond a single display-name for a channel and title+category+start/end+channel for a program. Even though Tribune isn't due to shut off service for another two months, I plan to have a zap2it scraper done by the weekend so that I don't have to worry about it. |
|
Possibly Related Threads… | |||||
Thread | Author | Replies | Views | Last Post | |
zap2it EPG missing detail in prime time | SamM | 1 | 401 |
2024-04-20, 02:13 PM Last Post: mvallevand |
|
zap2it xmltv different than listing | SamM | 4 | 1,516 |
2023-12-11, 02:42 PM Last Post: Prsa01 |
|
Anyone else having trouble with Zap2it? | ApexHack | 12 | 3,634 |
2008-02-13, 05:08 AM Last Post: ApexHack |
|
Another zap2it (and schedule direct) pay alternative | -Oz- | 0 | 3,306 |
2007-08-24, 03:03 PM Last Post: -Oz- |
|
Zap2it(i think this goes here) | dragonfiregum | 1 | 1,259 |
2007-04-07, 02:10 AM Last Post: Jim_ |
|
No zap2it expiry email | fla | 4 | 2,229 |
2007-02-12, 09:57 PM Last Post: fla |
|
Zap2it.com introduces TVlistings.com? | daphatty | 1 | 2,141 |
2006-11-29, 08:41 AM Last Post: DoPeFiSh |
|
Is zap2it free? | rlbond | 6 | 2,108 |
2006-09-14, 08:00 PM Last Post: Jim_ |
|
zap2it dilemna | PsychoRunner | 1 | 1,579 |
2006-07-29, 12:26 AM Last Post: sub |
|
Zap2it epg | Manoe | 2 | 1,885 |
2006-01-30, 06:18 PM Last Post: daphatty |