We require a small program to create a list of URLs ready for manual
submission to search engines.
The pages are contained within a local directory. There are between 20
and 70 pages per site. The url the pages are hosted on is within the
html page and will need to be parsed.
We need to be able to browse to a directory, and the program will make
a list like this
http://www.domain.com (extracted from a html template page which i
will upload for you to see) /pagename.htm (obtained from a file(s)
located in the directory)
eg
http://www.domain.com/pagename1.htm
http://www.domain.com/pagename2.htm
http://www.domain.com/pagename3.htm
http://www.domain.com/pagename4.htm
http://www.domain.com/pagename5.htm
http://www.domain.com/pagename6.htm
http://www.domain.com/pagename7.htm
http://www.domain.com/pagename8.htm
When we have this list, we can then copy and paste into the search
engine submission pages.
Please can you request clarification to ensure that i have explained
everything properly.
Many thanks |
Request for Question Clarification by
joseleon-ga
on
27 Nov 2003 01:25 PST
Hello, bmcompany:
Nice to see you again! I will try to do my best. This is what I
understand you want, a software that allows the following:
-Browse to a directory
-Read an html template located in that directory
-Parse that html template to extract the domain name
-Throw out a list of pages in that dir, fully qualyfied with the
domain name previously parsed
-This list can be copied to the clipboard (automatically if you want)
Is that ok?
Regards.
|
Clarification of Question by
bmcompany-ga
on
27 Nov 2003 02:53 PST
hi there, joseleon-ga. Nice to see you're still around, wasnt sure if
you would spot the post.
And the report analyser you made is working well and is being used
throughout the office!
You've got it right i think.
For each site, the pages have different text, images etc but is based
on the same template.
Download this zip file, it contains 5 sites. If you open one of the
htm files, you see a list of absolute links at the top. These links
contain the domain.
http://www.sesuk.net/ga/ga.zip
Please post back as soon as you have downloaded it and i'll take it down.
Also, could you output to a .txt file with the lists please.
Look forward to hearing from you soon.
|
Request for Question Clarification by
joseleon-ga
on
27 Nov 2003 03:07 PST
Hello, bmcompany:
You can safely remove the file, I have grabbed it. Nice to hear a
software by me is being used ;-).
Sure the results will be saved to a text file, I start working on it right now.
Regards.
|
Request for Question Clarification by
joseleon-ga
on
27 Nov 2003 03:18 PST
Hello, bmcompany:
Do you want me to list the "links contained into the files you sent
me" or to parse a "single" template file looking for a directory
domain and then add this domain to the "list of files contained into a
directory"?
For example, in the samples you sent me:
Executive_Transport.htm
Has this kind of links inside
href="http://www.limousinedirect.co.uk/Limousine_Hire.htm">Limousine
Service</a> | <a
href="http://www.limousinedirect.co.uk/Stretch_Limos.htm">St Andrews
Golf</a> | <a
.
.
.
Do you want to get that info? If that's the case, do you want to allow
duplicates, if any?
Regards.
|
Clarification of Question by
bmcompany-ga
on
27 Nov 2003 04:04 PST
Hi again,
The domain name will be the same for all the pages on each site, so in
one directory - there will be one domain. Therefore, you only need to
parse one of the pages - just to get the domain. The list is then
drawn from the page names in the dir, not the pages themselves.
We only link to a few of the pages inside the pages, therefore, the
list must be generated from pagenames in the dir, not the href links
in the page.
so, to recap. Simply parse one of the pages in the directory to get
the domain, then append the pagenames (from the dir) to the domain
that was stipped from the html.
Hope this is clear - let me know if you need further clarification
|
Request for Question Clarification by
joseleon-ga
on
27 Nov 2003 07:51 PST
Hello, bmcompany:
Please, download the first release from here:
http://www.xpde.com/LINKParser.zip
I have used the same interface that in the other software so you will
be familiar with it. Please, tell me if it does what you want.
Regards.
|
Clarification of Question by
bmcompany-ga
on
27 Nov 2003 08:27 PST
fantastic, seems to work fine.
Any chance of having an option to append the pages into a database for
an extra $50-$100?
say, Microsoft Access, table called tblpages. We can then add fields
for each search engine to say when we've done each submission.
Whenever the parser reads another list, its just adds records to the
database. However, we still need the txt file option.
Let me know what you think.
Thanks again - its looking great
|
Request for Question Clarification by
joseleon-ga
on
27 Nov 2003 11:01 PST
Hello, bmcompany:
Sure, no problem, you want also to export results to an Access
database table. Another option called Export to database, you choose
the MDB and the results are stored on it. Do you want to allow
duplicates? I will do this tomorrow morning GMT+1, is that ok for you?
Regards.
|
Clarification of Question by
bmcompany-ga
on
27 Nov 2003 13:16 PST
Firstly, to answer nivarty-ga's comment. I have seen literally dosens
of similar projects here on the Google Answers site, that is why i
started posting similar small projects here.
I have browsed both the FAQ's and the Terms of Service and can see
nothing that forbids this sort of question.
However, it is not my wish to violate any of Google's terms and if
someone at google requests that i remove the post, i will pay joseleon
for his time and without question will imediatly, and without
question, remove the post.
joseleon-ga, yes, tomorrow is fine - and the database shouldnt allow
duplicates. Thanks once again for your kind and prompt attention.
bmcompany
|
Clarification of Question by
bmcompany-ga
on
27 Nov 2003 13:32 PST
i have just sent an email to the GA support team, just to check that
this sort of question is allowed.
If this question is closed without me being notified, i will post
another question to pay you for your time so far.
Thanks again.
|
Request for Question Clarification by
joseleon-ga
on
28 Nov 2003 00:04 PST
Hello, bmcompany:
Thanks for your support, but I also asked some time ago to Google
support regarding this kind of questiones and they tell me are
allowed, no problem with them.
I will post today the database capable version.
Regards.
|