I am developer. i am developing a web application, in which, i need to display/embed docx file in div/web page part.
I have tried google api with using iframe code but not working like -
<iframe src="http://docs.google.com/gview?url=http://example.com/my.docx&embedded=true" style="width:600px; height:500px;" frameborder="0"></iframe>
but my application can not access by outside of my company. so this is not public documents. we can not upload the documents on google server.
So i am very confuse about the case ago a lot of time.
Can we dispaly docx file into web page ?
Please give a suggestion to resolve this issue.
Thanks.
As of just now, you can try https://nativedocuments.com/
Native Documents has a component you can embed to view or edit Word (doc/docx) files in a web app. You can host it, so your documents aren't transmitted over the public internet.
It is commercial software, but there are free tiers which may be enough for you.
Disclosure: I have an interest in Native Documents.
Related
I have a sharepoint at my office. Its 2013 version. Where I want to write some asp code. But the issue is SharePoint is blocking the code and I am getting error "Code blocks are not allowed in this file". I searched google and found several links to solve the issue by saying make some changes to the webconfig file.
Now my question is how do I find the file. Where it is actually.
What I have is a sharepoint, I don't have any designer. I only have admin access for this site. Can some one please guide me.
I know there are several entries here in stackoverflow, but no one is talking about where to find the file.
Please help me.
My apologies if this happens to be a repetition, in that case please point me to the right post. Thank you guys.
By default injecting server-side code (ASP.NET) in SharePoint pages directly from sites is not allowed for performance reasons, and should remain as is.
If you never approched SP developpment and are not an administrator of the farm in your company I strongly advise you to see first if you can solve your needs with client side development (javascript) instead of going to server side (ASP.NET).
SPS2013 comes with the "Script Editor WebPart" that you can use to inject your custom JS on pages. If you need your custom on all pages consider adding your JS on the site's masterpage.
From JS you can use SharePoint REST API to interact with your site https://learn.microsoft.com/en-us/sharepoint/dev/sp-add-ins/get-to-know-the-sharepoint-rest-service
If you need heavy customisation for your site you can move to the addin model (client side) that will require Visual Studio IDE develoment suite.
And last option is if you explicitly require serve side code and/or need to develop a scalable enterprise grade solution, you will need to make a "SharePoint full trust solution package".
PS: You may see articles around about "SharePoint Framework" (aka SPFx), unfortunatly this is not available for SPS2013.
I'm looking for a server for hosting images from a webservice that i'm working on. This webservice will need to access the images many times, I'll upload about 4GB of images per day to show to the users. My idea is to host the images over there and get the public links to put on the HTML.
So I'd like to know if Dropbox is an adequate tool for this, because I was studying the Dropbox API and I think It doesn't offer adequate tools to get the images's public link.
Summarizing my question, are these kind of hosts for this kind of services or not?
As the other comments and answers say, a normal CDN is probably a better choice for this.
For reference though, the Dropbox API does let you get publicly shareable links to files:
https://www.dropbox.com/developers/core/docs#shares
You can also modify these links as desired:
https://www.dropbox.com/help/201
There's also a direct and temporary version:
https://www.dropbox.com/developers/core/docs#media
Note however that there are bandwidth limits on these links:
https://www.dropbox.com/help/4204
Also, the API enables you to access file content directly:
https://www.dropbox.com/developers/core/docs#files-GET
Try amazon web services http://aws.amazon.com/s3/
Lots of big websites use this for serving images. It is very fast!
Is there a JavaScript API for Office Web Apps (specifically - Word Web App) that I can use in order to get notifications from the Word doc?
My scenario is this:
I have an ASP.NET application, in which I want to display a word document using the Word Web App (on premises). I would like to embed the document in the web page using iFrame (haven't tried it yet, hope it'll work...).
I would like to get notification when the user saved the document, so that when this happens, I can remove the iFrame and display something else.
Is that possible?
I recently read MSDN article on this - http://msdn.microsoft.com/en-us/magazine/jj891051.aspx. Hope that helps.
I made an ASP.NET MVC application which allows user to create dynamic websites. I need to add feature which will allow to download from server off-line version of choosen website as static html files with menu, hyperlinks, images, documents etc. It should work similar to applications such as Teleport Pro, but I have to choose from Admin Panel which content should be export.
Client wants to burn static website on CD, save on pendrive.
Do you have any ideas how to begin? Please help.
I currently have implemented that in a current project...
User is able to change anything in the frontend and at the end he can publish and download the offline files... the site subscribe users and show all prizes, winners and more information about that campaign.
All was done in ASP.NET MVC3 under .NET4 and hosted in AppHarbor.
It's composed at several applications but for what you want, you develop the Backend and the Frontend, and to generate the static files, simple use the Frontend to grab the full HTML
As an example, I can show what 2 users did...
Callme.dk did http://callme.julekal.info and
Sony Nordic did http://sony.julekal.info
plus, you can simply point custom domains to it as well like http://sonynordicxmas.net/
To publish and generate all files:
one part of the editing:
So I give the users, offline access (through the .zip file), online access (through the frontend application) and the ability of using custom domains...
I think the only way this might be possible is if you go to every single page and then use your browser to "Save" the web page script and all.
However this causes several issues;
You never quite get everything and you need to massage the HTML produced, dowload all the images etc to get the page to look right
Each html file now has an associated folder with the same name and each time you do this you will get another html file with a folder. You can combine all the folders into a single one but that leads me to item 3.
You will need to edit each html file to clear up any pathing issues if you want to share a single source folder.
Data is no longer dynamic!
You need to, if you want to link all the pages to each other, edit every single html file and resolver the anchor tags.
This is too much work and I think it actually breaks the true requirement.
Don't do it! :)
I'm asking on behalf of somebody, so I don't have too many details.
What options are available for indexing site content in an ASP.NET web site? I suspect SQL Server's Full Text index may be used if the page content is stored in the database. How would I index dynamic and static content if that content isn't stored in the DB, but in html and aspx pages themselves?
We purchased Karamasoft Ultimate Search several years ago. It is a search engine add-on for your web site. I like it because it is a simple tool that taught us searching on our site. It is pretty inexpensive and we knew we could buy later if we needed more or different features. We needed something that would give us searching without having to do a lot of programming.
Specifically, this tool is a web crawler. It will run on your web server and it will act like an end-user and navigate through your site keeping a record of your web pages, so when a real users searches, they are told the pages that have the content they want.
Keep that in mind it is acting like an end-user, so your dynamic data is indexed right along with the static stuff because it indexes the final web page. We needed this feature and it is what appealed to us the most.
You can use a web crawler to crawl that site and add the content to a database which then is full text indexed. There are a number of web crawlers out there.
Lucene is a well known open source tool that would help you here. The main branch is Java based but there is a .Net port too.
Main site: http://lucene.apache.org/
.Net port: http://incubator.apache.org/lucene.net/
Having used several alternatives I would be loath to do anything other than Google Site Search.
The only reason I use SQL Full Text Search is to search through multiple columns. It's really hard to implement it in any effective manner.