Google Answers Logo
View Question
 
Q: for joseleon-ga only please ( Answered 5 out of 5 stars,   0 Comments )
Question  
Subject: for joseleon-ga only please
Category: Computers
Asked by: bmcompany-ga
List Price: $150.00
Posted: 26 Nov 2003 07:05 PST
Expires: 26 Dec 2003 07:05 PST
Question ID: 280786
We require a small program to create a list of URLs ready for manual
submission to search engines.

The pages are contained within a local directory. There are between 20
and 70 pages per site. The url the pages are hosted on is within the
html page and will need to be parsed.

We need to be able to browse to a directory, and the program will make
a list like this

http://www.domain.com (extracted from a html template page which i
will upload for you to see) /pagename.htm (obtained from a file(s)
located in the directory)

eg

http://www.domain.com/pagename1.htm
http://www.domain.com/pagename2.htm
http://www.domain.com/pagename3.htm
http://www.domain.com/pagename4.htm
http://www.domain.com/pagename5.htm
http://www.domain.com/pagename6.htm
http://www.domain.com/pagename7.htm
http://www.domain.com/pagename8.htm

When we have this list, we can then copy and paste into the search
engine submission pages.

Please can you request clarification to ensure that i have explained
everything properly.

Many thanks

Request for Question Clarification by joseleon-ga on 27 Nov 2003 01:25 PST
Hello, bmcompany:
  Nice to see you again! I will try to do my best. This is what I
understand you want, a software that allows the following:
-Browse to a directory
-Read an html template located in that directory
-Parse that html template to extract the domain name
-Throw out a list of pages in that dir, fully qualyfied with the
domain name previously parsed
-This list can be copied to the clipboard (automatically if you want)

Is that ok?

Regards.

Clarification of Question by bmcompany-ga on 27 Nov 2003 02:53 PST
hi there, joseleon-ga. Nice to see you're still around, wasnt sure if
you would spot the post.

And the report analyser you made is working well and is being used
throughout the office!

You've got it right i think.

For each site, the pages have different text, images etc but is based
on the same template.

Download this zip file, it contains 5 sites. If you open one of the
htm files, you see a list of absolute links at the top. These links
contain the domain.

http://www.sesuk.net/ga/ga.zip

Please post back as soon as you have downloaded it and i'll take it down.

Also, could you output to a .txt file with the lists please.

Look forward to hearing from you soon.

Request for Question Clarification by joseleon-ga on 27 Nov 2003 03:07 PST
Hello, bmcompany:
  You can safely remove the file, I have grabbed it. Nice to hear a
software by me is being used ;-).

  Sure the results will be saved to a text file, I start working on it right now.

Regards.

Request for Question Clarification by joseleon-ga on 27 Nov 2003 03:18 PST
Hello, bmcompany:
  Do you want me to list the "links contained into the files you sent
me" or to parse a "single" template file looking for a directory
domain and then add this domain to the "list of files contained into a
directory"?
  
For example, in the samples you sent me:

Executive_Transport.htm

Has this kind of links inside

href="http://www.limousinedirect.co.uk/Limousine_Hire.htm">Limousine
Service</a>&nbsp;|&nbsp;<a
href="http://www.limousinedirect.co.uk/Stretch_Limos.htm">St Andrews
Golf</a>&nbsp;|&nbsp;<a
.
.
.

Do you want to get that info? If that's the case, do you want to allow
duplicates, if any?

Regards.

Clarification of Question by bmcompany-ga on 27 Nov 2003 04:04 PST
Hi again,

The domain name will be the same for all the pages on each site, so in
one directory - there will be one domain. Therefore, you only need to
parse one of the pages - just to get the domain. The list is then
drawn from the page names in the dir, not the pages themselves.

We only link to a few of the pages inside the pages, therefore, the
list must be generated from pagenames in the dir, not the href links
in the page.

so, to recap. Simply parse one of the pages in the directory to get
the domain, then append the pagenames (from the dir) to the domain
that was stipped from the html.

Hope this is clear - let me know if you need further clarification

Request for Question Clarification by joseleon-ga on 27 Nov 2003 07:51 PST
Hello, bmcompany:
  Please, download the first release from here:

http://www.xpde.com/LINKParser.zip

I have used the same interface that in the other software so you will
be familiar with it. Please, tell me if it does what you want.

Regards.

Clarification of Question by bmcompany-ga on 27 Nov 2003 08:27 PST
fantastic, seems to work fine.

Any chance of having an option to append the pages into a database for
an extra $50-$100?

say, Microsoft Access, table called tblpages. We can then add fields
for each search engine to say when we've done each submission.
Whenever the parser reads another list, its just adds records to the
database. However, we still need the txt file option.

Let me know what you think.

Thanks again - its looking great

Request for Question Clarification by joseleon-ga on 27 Nov 2003 11:01 PST
Hello, bmcompany:
  Sure, no problem, you want also to export results to an Access
database table. Another option called Export to database, you choose
the MDB and the results are stored on it. Do you want to allow
duplicates? I will do this tomorrow morning GMT+1, is that ok for you?

Regards.

Clarification of Question by bmcompany-ga on 27 Nov 2003 13:16 PST
Firstly, to answer nivarty-ga's comment. I have seen literally dosens
of similar projects here on the Google Answers site, that is why i
started posting similar small projects here.

I have browsed both the FAQ's and the Terms of Service and can see
nothing that forbids this sort of question.

However, it is not my wish to violate any of Google's terms and if
someone at google requests that i remove the post, i will pay joseleon
for his time and without question will imediatly, and without
question, remove the post.

joseleon-ga, yes, tomorrow is fine - and the database shouldnt allow
duplicates. Thanks once again for your kind and prompt attention.

bmcompany

Clarification of Question by bmcompany-ga on 27 Nov 2003 13:32 PST
i have just sent an email to the GA support team, just to check that
this sort of question is allowed.

If this question is closed without me being notified, i will post
another question to pay you for your time so far.

Thanks again.

Request for Question Clarification by joseleon-ga on 28 Nov 2003 00:04 PST
Hello, bmcompany:
  Thanks for your support, but I also asked some time ago to Google
support regarding this kind of questiones and they tell me are
allowed, no problem with them.

I will post today the database capable version.

Regards.
Answer  
Subject: Re: for joseleon-ga only please
Answered By: joseleon-ga on 28 Nov 2003 00:50 PST
Rated:5 out of 5 stars
 
Hello, bmcompany:
  
  I have finished the last version with database export capabilities,
you can find it here:

http://www.xpde.com/Link_parser.zip

Once you parse the files to get the results, the action of Database
Export will be enabled, then you will be requested for an access
database (*.mdb) and results will be saved there.

When the process is over, you get a small report of how many records
were added and how many were duplicated. In the zip there is a sample
access database you can use, bear in mind that already contains data.

You can use any access database as long as the table is called
"link_stats" and the field to insert the links is called "link", no
matter the position of the field.

I hope this is what you were looking for, and as always, don't
hesitate to request for any clarification.

Regards.

Clarification of Answer by joseleon-ga on 28 Nov 2003 01:17 PST
Hello, bmcompany:
 Thanks a lot for the great tip and for your requests ;-)

Regards.
bmcompany-ga rated this answer:5 out of 5 stars and gave an additional tip of: $100.00
Perfect (as always)

keep your eyes open for posts in the near future.

Comments  
There are no comments at this time.

Important Disclaimer: Answers and comments provided on Google Answers are general information, and are not intended to substitute for informed professional medical, psychiatric, psychological, tax, legal, investment, accounting, or other professional advice. Google does not endorse, and expressly disclaims liability for any product, manufacturer, distributor, service or service provider mentioned or any opinion expressed in answers or comments. Please read carefully the Google Answers Terms of Service.

If you feel that you have found inappropriate content, please let us know by emailing us at answers-support@google.com with the question ID listed above. Thank you.
Search Google Answers for
Google Answers  


Google Home - Answers FAQ - Terms of Service - Privacy Policy